00:00:00.001 Started by upstream project "autotest-spdk-master-vs-dpdk-v22.11" build number 2435 00:00:00.001 originally caused by: 00:00:00.001 Started by upstream project "nightly-trigger" build number 3700 00:00:00.001 originally caused by: 00:00:00.001 Started by timer 00:00:00.008 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-vg.groovy 00:00:00.009 The recommended git tool is: git 00:00:00.009 using credential 00000000-0000-0000-0000-000000000002 00:00:00.011 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.021 Fetching changes from the remote Git repository 00:00:00.025 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.038 Using shallow fetch with depth 1 00:00:00.038 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.038 > git --version # timeout=10 00:00:00.052 > git --version # 'git version 2.39.2' 00:00:00.052 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.070 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.070 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:02.393 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:02.406 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:02.417 Checking out Revision db4637e8b949f278f369ec13f70585206ccd9507 (FETCH_HEAD) 00:00:02.417 > git config core.sparsecheckout # timeout=10 00:00:02.429 > git read-tree -mu HEAD # timeout=10 00:00:02.445 > git checkout -f db4637e8b949f278f369ec13f70585206ccd9507 # timeout=5 00:00:02.470 Commit message: "jenkins/jjb-config: Add missing SPDK_TEST_NVME_INTERRUPT flag" 00:00:02.471 > git rev-list --no-walk db4637e8b949f278f369ec13f70585206ccd9507 # timeout=10 00:00:02.704 [Pipeline] Start of Pipeline 00:00:02.720 [Pipeline] library 00:00:02.722 Loading library shm_lib@master 00:00:02.722 Library shm_lib@master is cached. Copying from home. 00:00:02.740 [Pipeline] node 00:00:02.762 Running on VM-host-WFP7 in /var/jenkins/workspace/raid-vg-autotest 00:00:02.763 [Pipeline] { 00:00:02.773 [Pipeline] catchError 00:00:02.774 [Pipeline] { 00:00:02.784 [Pipeline] wrap 00:00:02.791 [Pipeline] { 00:00:02.797 [Pipeline] stage 00:00:02.798 [Pipeline] { (Prologue) 00:00:02.811 [Pipeline] echo 00:00:02.812 Node: VM-host-WFP7 00:00:02.816 [Pipeline] cleanWs 00:00:02.826 [WS-CLEANUP] Deleting project workspace... 00:00:02.826 [WS-CLEANUP] Deferred wipeout is used... 00:00:02.832 [WS-CLEANUP] done 00:00:03.047 [Pipeline] setCustomBuildProperty 00:00:03.137 [Pipeline] httpRequest 00:00:03.822 [Pipeline] echo 00:00:03.824 Sorcerer 10.211.164.20 is alive 00:00:03.834 [Pipeline] retry 00:00:03.836 [Pipeline] { 00:00:03.849 [Pipeline] httpRequest 00:00:03.854 HttpMethod: GET 00:00:03.855 URL: http://10.211.164.20/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:03.855 Sending request to url: http://10.211.164.20/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:03.863 Response Code: HTTP/1.1 200 OK 00:00:03.863 Success: Status code 200 is in the accepted range: 200,404 00:00:03.864 Saving response body to /var/jenkins/workspace/raid-vg-autotest/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:04.593 [Pipeline] } 00:00:04.605 [Pipeline] // retry 00:00:04.611 [Pipeline] sh 00:00:04.890 + tar --no-same-owner -xf jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:04.905 [Pipeline] httpRequest 00:00:05.246 [Pipeline] echo 00:00:05.248 Sorcerer 10.211.164.20 is alive 00:00:05.254 [Pipeline] retry 00:00:05.256 [Pipeline] { 00:00:05.266 [Pipeline] httpRequest 00:00:05.271 HttpMethod: GET 00:00:05.272 URL: http://10.211.164.20/packages/spdk_8d3947977640da882a3cdcc21a7575115b7e7787.tar.gz 00:00:05.272 Sending request to url: http://10.211.164.20/packages/spdk_8d3947977640da882a3cdcc21a7575115b7e7787.tar.gz 00:00:05.274 Response Code: HTTP/1.1 200 OK 00:00:05.275 Success: Status code 200 is in the accepted range: 200,404 00:00:05.275 Saving response body to /var/jenkins/workspace/raid-vg-autotest/spdk_8d3947977640da882a3cdcc21a7575115b7e7787.tar.gz 00:00:30.581 [Pipeline] } 00:00:30.600 [Pipeline] // retry 00:00:30.610 [Pipeline] sh 00:00:30.897 + tar --no-same-owner -xf spdk_8d3947977640da882a3cdcc21a7575115b7e7787.tar.gz 00:00:33.452 [Pipeline] sh 00:00:33.739 + git -C spdk log --oneline -n5 00:00:33.739 8d3947977 spdk_dd: simplify `io_uring_peek_cqe` return code processing 00:00:33.739 77ee034c7 bdev/nvme: Add lock to unprotected operations around attach controller 00:00:33.739 48454bb28 bdev/nvme: Add lock to unprotected operations around detach controller 00:00:33.739 4b59d7893 bdev/nvme: Use nbdev always for local nvme_bdev pointer variables 00:00:33.739 e56f1618f lib/ftl: Add explicit support for write unit sizes of base device 00:00:33.761 [Pipeline] withCredentials 00:00:33.774 > git --version # timeout=10 00:00:33.790 > git --version # 'git version 2.39.2' 00:00:33.809 Masking supported pattern matches of $GIT_PASSWORD or $GIT_ASKPASS 00:00:33.811 [Pipeline] { 00:00:33.821 [Pipeline] retry 00:00:33.823 [Pipeline] { 00:00:33.840 [Pipeline] sh 00:00:34.126 + git ls-remote http://dpdk.org/git/dpdk-stable v22.11.4 00:00:34.399 [Pipeline] } 00:00:34.418 [Pipeline] // retry 00:00:34.424 [Pipeline] } 00:00:34.441 [Pipeline] // withCredentials 00:00:34.452 [Pipeline] httpRequest 00:00:34.840 [Pipeline] echo 00:00:34.843 Sorcerer 10.211.164.20 is alive 00:00:34.855 [Pipeline] retry 00:00:34.858 [Pipeline] { 00:00:34.874 [Pipeline] httpRequest 00:00:34.879 HttpMethod: GET 00:00:34.880 URL: http://10.211.164.20/packages/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:00:34.880 Sending request to url: http://10.211.164.20/packages/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:00:34.893 Response Code: HTTP/1.1 200 OK 00:00:34.894 Success: Status code 200 is in the accepted range: 200,404 00:00:34.894 Saving response body to /var/jenkins/workspace/raid-vg-autotest/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:01:29.277 [Pipeline] } 00:01:29.294 [Pipeline] // retry 00:01:29.303 [Pipeline] sh 00:01:29.587 + tar --no-same-owner -xf dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:01:30.980 [Pipeline] sh 00:01:31.266 + git -C dpdk log --oneline -n5 00:01:31.266 caf0f5d395 version: 22.11.4 00:01:31.266 7d6f1cc05f Revert "net/iavf: fix abnormal disable HW interrupt" 00:01:31.266 dc9c799c7d vhost: fix missing spinlock unlock 00:01:31.266 4307659a90 net/mlx5: fix LACP redirection in Rx domain 00:01:31.266 6ef77f2a5e net/gve: fix RX buffer size alignment 00:01:31.286 [Pipeline] writeFile 00:01:31.301 [Pipeline] sh 00:01:31.588 + jbp/jenkins/jjb-config/jobs/scripts/autorun_quirks.sh 00:01:31.600 [Pipeline] sh 00:01:31.889 + cat autorun-spdk.conf 00:01:31.889 SPDK_RUN_FUNCTIONAL_TEST=1 00:01:31.889 SPDK_RUN_ASAN=1 00:01:31.889 SPDK_RUN_UBSAN=1 00:01:31.889 SPDK_TEST_RAID=1 00:01:31.889 SPDK_TEST_NATIVE_DPDK=v22.11.4 00:01:31.889 SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:31.889 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:31.906 RUN_NIGHTLY=1 00:01:31.907 [Pipeline] } 00:01:31.918 [Pipeline] // stage 00:01:31.930 [Pipeline] stage 00:01:31.932 [Pipeline] { (Run VM) 00:01:31.941 [Pipeline] sh 00:01:32.221 + jbp/jenkins/jjb-config/jobs/scripts/prepare_nvme.sh 00:01:32.221 + echo 'Start stage prepare_nvme.sh' 00:01:32.221 Start stage prepare_nvme.sh 00:01:32.221 + [[ -n 0 ]] 00:01:32.221 + disk_prefix=ex0 00:01:32.221 + [[ -n /var/jenkins/workspace/raid-vg-autotest ]] 00:01:32.221 + [[ -e /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf ]] 00:01:32.221 + source /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf 00:01:32.221 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:01:32.221 ++ SPDK_RUN_ASAN=1 00:01:32.221 ++ SPDK_RUN_UBSAN=1 00:01:32.221 ++ SPDK_TEST_RAID=1 00:01:32.221 ++ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:01:32.221 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:32.221 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:32.221 ++ RUN_NIGHTLY=1 00:01:32.221 + cd /var/jenkins/workspace/raid-vg-autotest 00:01:32.221 + nvme_files=() 00:01:32.221 + declare -A nvme_files 00:01:32.221 + backend_dir=/var/lib/libvirt/images/backends 00:01:32.221 + nvme_files['nvme.img']=5G 00:01:32.221 + nvme_files['nvme-cmb.img']=5G 00:01:32.221 + nvme_files['nvme-multi0.img']=4G 00:01:32.221 + nvme_files['nvme-multi1.img']=4G 00:01:32.221 + nvme_files['nvme-multi2.img']=4G 00:01:32.221 + nvme_files['nvme-openstack.img']=8G 00:01:32.221 + nvme_files['nvme-zns.img']=5G 00:01:32.221 + (( SPDK_TEST_NVME_PMR == 1 )) 00:01:32.221 + (( SPDK_TEST_FTL == 1 )) 00:01:32.221 + (( SPDK_TEST_NVME_FDP == 1 )) 00:01:32.221 + [[ ! -d /var/lib/libvirt/images/backends ]] 00:01:32.221 + for nvme in "${!nvme_files[@]}" 00:01:32.221 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-multi2.img -s 4G 00:01:32.221 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-multi2.img', fmt=raw size=4294967296 preallocation=falloc 00:01:32.221 + for nvme in "${!nvme_files[@]}" 00:01:32.221 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-cmb.img -s 5G 00:01:32.221 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-cmb.img', fmt=raw size=5368709120 preallocation=falloc 00:01:32.221 + for nvme in "${!nvme_files[@]}" 00:01:32.221 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-openstack.img -s 8G 00:01:32.221 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-openstack.img', fmt=raw size=8589934592 preallocation=falloc 00:01:32.221 + for nvme in "${!nvme_files[@]}" 00:01:32.221 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-zns.img -s 5G 00:01:32.221 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-zns.img', fmt=raw size=5368709120 preallocation=falloc 00:01:32.221 + for nvme in "${!nvme_files[@]}" 00:01:32.221 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-multi1.img -s 4G 00:01:32.221 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-multi1.img', fmt=raw size=4294967296 preallocation=falloc 00:01:32.221 + for nvme in "${!nvme_files[@]}" 00:01:32.221 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme-multi0.img -s 4G 00:01:32.221 Formatting '/var/lib/libvirt/images/backends/ex0-nvme-multi0.img', fmt=raw size=4294967296 preallocation=falloc 00:01:32.221 + for nvme in "${!nvme_files[@]}" 00:01:32.221 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex0-nvme.img -s 5G 00:01:32.480 Formatting '/var/lib/libvirt/images/backends/ex0-nvme.img', fmt=raw size=5368709120 preallocation=falloc 00:01:32.481 ++ sudo grep -rl ex0-nvme.img /etc/libvirt/qemu 00:01:32.481 + echo 'End stage prepare_nvme.sh' 00:01:32.481 End stage prepare_nvme.sh 00:01:32.493 [Pipeline] sh 00:01:32.779 + DISTRO=fedora39 CPUS=10 RAM=12288 jbp/jenkins/jjb-config/jobs/scripts/vagrant_create_vm.sh 00:01:32.779 Setup: -n 10 -s 12288 -x http://proxy-dmz.intel.com:911 -p libvirt --qemu-emulator=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 -b /var/lib/libvirt/images/backends/ex0-nvme.img -b /var/lib/libvirt/images/backends/ex0-nvme-multi0.img,nvme,/var/lib/libvirt/images/backends/ex0-nvme-multi1.img:/var/lib/libvirt/images/backends/ex0-nvme-multi2.img -H -a -v -f fedora39 00:01:32.779 00:01:32.779 DIR=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant 00:01:32.779 SPDK_DIR=/var/jenkins/workspace/raid-vg-autotest/spdk 00:01:32.779 VAGRANT_TARGET=/var/jenkins/workspace/raid-vg-autotest 00:01:32.779 HELP=0 00:01:32.779 DRY_RUN=0 00:01:32.779 NVME_FILE=/var/lib/libvirt/images/backends/ex0-nvme.img,/var/lib/libvirt/images/backends/ex0-nvme-multi0.img, 00:01:32.779 NVME_DISKS_TYPE=nvme,nvme, 00:01:32.779 NVME_AUTO_CREATE=0 00:01:32.779 NVME_DISKS_NAMESPACES=,/var/lib/libvirt/images/backends/ex0-nvme-multi1.img:/var/lib/libvirt/images/backends/ex0-nvme-multi2.img, 00:01:32.779 NVME_CMB=,, 00:01:32.779 NVME_PMR=,, 00:01:32.779 NVME_ZNS=,, 00:01:32.779 NVME_MS=,, 00:01:32.779 NVME_FDP=,, 00:01:32.779 SPDK_VAGRANT_DISTRO=fedora39 00:01:32.779 SPDK_VAGRANT_VMCPU=10 00:01:32.779 SPDK_VAGRANT_VMRAM=12288 00:01:32.779 SPDK_VAGRANT_PROVIDER=libvirt 00:01:32.779 SPDK_VAGRANT_HTTP_PROXY=http://proxy-dmz.intel.com:911 00:01:32.779 SPDK_QEMU_EMULATOR=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 00:01:32.779 SPDK_OPENSTACK_NETWORK=0 00:01:32.779 VAGRANT_PACKAGE_BOX=0 00:01:32.779 VAGRANTFILE=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant/Vagrantfile 00:01:32.779 FORCE_DISTRO=true 00:01:32.780 VAGRANT_BOX_VERSION= 00:01:32.780 EXTRA_VAGRANTFILES= 00:01:32.780 NIC_MODEL=virtio 00:01:32.780 00:01:32.780 mkdir: created directory '/var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt' 00:01:32.780 /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt /var/jenkins/workspace/raid-vg-autotest 00:01:34.688 Bringing machine 'default' up with 'libvirt' provider... 00:01:34.948 ==> default: Creating image (snapshot of base box volume). 00:01:35.208 ==> default: Creating domain with the following settings... 00:01:35.208 ==> default: -- Name: fedora39-39-1.5-1721788873-2326_default_1733424712_c7c9ed286b37618ae4c5 00:01:35.208 ==> default: -- Domain type: kvm 00:01:35.208 ==> default: -- Cpus: 10 00:01:35.208 ==> default: -- Feature: acpi 00:01:35.208 ==> default: -- Feature: apic 00:01:35.208 ==> default: -- Feature: pae 00:01:35.208 ==> default: -- Memory: 12288M 00:01:35.208 ==> default: -- Memory Backing: hugepages: 00:01:35.208 ==> default: -- Management MAC: 00:01:35.208 ==> default: -- Loader: 00:01:35.208 ==> default: -- Nvram: 00:01:35.208 ==> default: -- Base box: spdk/fedora39 00:01:35.209 ==> default: -- Storage pool: default 00:01:35.209 ==> default: -- Image: /var/lib/libvirt/images/fedora39-39-1.5-1721788873-2326_default_1733424712_c7c9ed286b37618ae4c5.img (20G) 00:01:35.209 ==> default: -- Volume Cache: default 00:01:35.209 ==> default: -- Kernel: 00:01:35.209 ==> default: -- Initrd: 00:01:35.209 ==> default: -- Graphics Type: vnc 00:01:35.209 ==> default: -- Graphics Port: -1 00:01:35.209 ==> default: -- Graphics IP: 127.0.0.1 00:01:35.209 ==> default: -- Graphics Password: Not defined 00:01:35.209 ==> default: -- Video Type: cirrus 00:01:35.209 ==> default: -- Video VRAM: 9216 00:01:35.209 ==> default: -- Sound Type: 00:01:35.209 ==> default: -- Keymap: en-us 00:01:35.209 ==> default: -- TPM Path: 00:01:35.209 ==> default: -- INPUT: type=mouse, bus=ps2 00:01:35.209 ==> default: -- Command line args: 00:01:35.209 ==> default: -> value=-device, 00:01:35.209 ==> default: -> value=nvme,id=nvme-0,serial=12340,addr=0x10, 00:01:35.209 ==> default: -> value=-drive, 00:01:35.209 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme.img,if=none,id=nvme-0-drive0, 00:01:35.209 ==> default: -> value=-device, 00:01:35.209 ==> default: -> value=nvme-ns,drive=nvme-0-drive0,bus=nvme-0,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:35.209 ==> default: -> value=-device, 00:01:35.209 ==> default: -> value=nvme,id=nvme-1,serial=12341,addr=0x11, 00:01:35.209 ==> default: -> value=-drive, 00:01:35.209 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-multi0.img,if=none,id=nvme-1-drive0, 00:01:35.209 ==> default: -> value=-device, 00:01:35.209 ==> default: -> value=nvme-ns,drive=nvme-1-drive0,bus=nvme-1,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:35.209 ==> default: -> value=-drive, 00:01:35.209 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-multi1.img,if=none,id=nvme-1-drive1, 00:01:35.209 ==> default: -> value=-device, 00:01:35.209 ==> default: -> value=nvme-ns,drive=nvme-1-drive1,bus=nvme-1,nsid=2,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:35.209 ==> default: -> value=-drive, 00:01:35.209 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-multi2.img,if=none,id=nvme-1-drive2, 00:01:35.209 ==> default: -> value=-device, 00:01:35.209 ==> default: -> value=nvme-ns,drive=nvme-1-drive2,bus=nvme-1,nsid=3,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:35.209 ==> default: Creating shared folders metadata... 00:01:35.209 ==> default: Starting domain. 00:01:36.590 ==> default: Waiting for domain to get an IP address... 00:01:58.538 ==> default: Waiting for SSH to become available... 00:01:58.538 ==> default: Configuring and enabling network interfaces... 00:02:02.750 default: SSH address: 192.168.121.28:22 00:02:02.750 default: SSH username: vagrant 00:02:02.750 default: SSH auth method: private key 00:02:06.071 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/spdk/ => /home/vagrant/spdk_repo/spdk 00:02:14.196 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/dpdk/ => /home/vagrant/spdk_repo/dpdk 00:02:19.498 ==> default: Mounting SSHFS shared folder... 00:02:22.038 ==> default: Mounting folder via SSHFS: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/fedora39-libvirt/output => /home/vagrant/spdk_repo/output 00:02:22.038 ==> default: Checking Mount.. 00:02:23.946 ==> default: Folder Successfully Mounted! 00:02:23.946 ==> default: Running provisioner: file... 00:02:24.889 default: ~/.gitconfig => .gitconfig 00:02:25.497 00:02:25.497 SUCCESS! 00:02:25.497 00:02:25.497 cd to /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt and type "vagrant ssh" to use. 00:02:25.497 Use vagrant "suspend" and vagrant "resume" to stop and start. 00:02:25.497 Use vagrant "destroy" followed by "rm -rf /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt" to destroy all trace of vm. 00:02:25.497 00:02:25.506 [Pipeline] } 00:02:25.520 [Pipeline] // stage 00:02:25.530 [Pipeline] dir 00:02:25.530 Running in /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt 00:02:25.532 [Pipeline] { 00:02:25.543 [Pipeline] catchError 00:02:25.544 [Pipeline] { 00:02:25.556 [Pipeline] sh 00:02:25.838 + vagrant ssh-config --host vagrant 00:02:25.838 + sed -ne /^Host/,$p 00:02:25.838 + tee ssh_conf 00:02:28.373 Host vagrant 00:02:28.373 HostName 192.168.121.28 00:02:28.373 User vagrant 00:02:28.373 Port 22 00:02:28.373 UserKnownHostsFile /dev/null 00:02:28.373 StrictHostKeyChecking no 00:02:28.373 PasswordAuthentication no 00:02:28.373 IdentityFile /var/lib/libvirt/images/.vagrant.d/boxes/spdk-VAGRANTSLASH-fedora39/39-1.5-1721788873-2326/libvirt/fedora39 00:02:28.373 IdentitiesOnly yes 00:02:28.373 LogLevel FATAL 00:02:28.373 ForwardAgent yes 00:02:28.373 ForwardX11 yes 00:02:28.373 00:02:28.387 [Pipeline] withEnv 00:02:28.389 [Pipeline] { 00:02:28.401 [Pipeline] sh 00:02:28.740 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant #!/bin/bash 00:02:28.740 source /etc/os-release 00:02:28.740 [[ -e /image.version ]] && img=$(< /image.version) 00:02:28.740 # Minimal, systemd-like check. 00:02:28.740 if [[ -e /.dockerenv ]]; then 00:02:28.740 # Clear garbage from the node's name: 00:02:28.740 # agt-er_autotest_547-896 -> autotest_547-896 00:02:28.740 # $HOSTNAME is the actual container id 00:02:28.740 agent=$HOSTNAME@${DOCKER_SWARM_PLUGIN_JENKINS_AGENT_NAME#*_} 00:02:28.740 if grep -q "/etc/hostname" /proc/self/mountinfo; then 00:02:28.740 # We can assume this is a mount from a host where container is running, 00:02:28.740 # so fetch its hostname to easily identify the target swarm worker. 00:02:28.740 container="$(< /etc/hostname) ($agent)" 00:02:28.740 else 00:02:28.740 # Fallback 00:02:28.740 container=$agent 00:02:28.740 fi 00:02:28.740 fi 00:02:28.740 echo "${NAME} ${VERSION_ID}|$(uname -r)|${img:-N/A}|${container:-N/A}" 00:02:28.740 00:02:29.011 [Pipeline] } 00:02:29.026 [Pipeline] // withEnv 00:02:29.033 [Pipeline] setCustomBuildProperty 00:02:29.047 [Pipeline] stage 00:02:29.049 [Pipeline] { (Tests) 00:02:29.065 [Pipeline] sh 00:02:29.346 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh vagrant@vagrant:./ 00:02:29.619 [Pipeline] sh 00:02:29.906 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/pkgdep-autoruner.sh vagrant@vagrant:./ 00:02:30.178 [Pipeline] timeout 00:02:30.179 Timeout set to expire in 1 hr 30 min 00:02:30.180 [Pipeline] { 00:02:30.193 [Pipeline] sh 00:02:30.477 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant git -C spdk_repo/spdk reset --hard 00:02:31.049 HEAD is now at 8d3947977 spdk_dd: simplify `io_uring_peek_cqe` return code processing 00:02:31.062 [Pipeline] sh 00:02:31.344 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant sudo chown vagrant:vagrant spdk_repo 00:02:31.615 [Pipeline] sh 00:02:31.900 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf vagrant@vagrant:spdk_repo 00:02:32.176 [Pipeline] sh 00:02:32.463 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant JOB_BASE_NAME=raid-vg-autotest ./autoruner.sh spdk_repo 00:02:32.723 ++ readlink -f spdk_repo 00:02:32.723 + DIR_ROOT=/home/vagrant/spdk_repo 00:02:32.723 + [[ -n /home/vagrant/spdk_repo ]] 00:02:32.723 + DIR_SPDK=/home/vagrant/spdk_repo/spdk 00:02:32.723 + DIR_OUTPUT=/home/vagrant/spdk_repo/output 00:02:32.723 + [[ -d /home/vagrant/spdk_repo/spdk ]] 00:02:32.723 + [[ ! -d /home/vagrant/spdk_repo/output ]] 00:02:32.723 + [[ -d /home/vagrant/spdk_repo/output ]] 00:02:32.723 + [[ raid-vg-autotest == pkgdep-* ]] 00:02:32.723 + cd /home/vagrant/spdk_repo 00:02:32.723 + source /etc/os-release 00:02:32.723 ++ NAME='Fedora Linux' 00:02:32.723 ++ VERSION='39 (Cloud Edition)' 00:02:32.723 ++ ID=fedora 00:02:32.723 ++ VERSION_ID=39 00:02:32.723 ++ VERSION_CODENAME= 00:02:32.723 ++ PLATFORM_ID=platform:f39 00:02:32.723 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:02:32.723 ++ ANSI_COLOR='0;38;2;60;110;180' 00:02:32.723 ++ LOGO=fedora-logo-icon 00:02:32.723 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:02:32.723 ++ HOME_URL=https://fedoraproject.org/ 00:02:32.723 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:02:32.723 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:02:32.723 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:02:32.723 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:02:32.723 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:02:32.723 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:02:32.723 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:02:32.723 ++ SUPPORT_END=2024-11-12 00:02:32.723 ++ VARIANT='Cloud Edition' 00:02:32.723 ++ VARIANT_ID=cloud 00:02:32.723 + uname -a 00:02:32.723 Linux fedora39-cloud-1721788873-2326 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:02:32.723 + sudo /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:02:33.292 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:02:33.292 Hugepages 00:02:33.292 node hugesize free / total 00:02:33.292 node0 1048576kB 0 / 0 00:02:33.292 node0 2048kB 0 / 0 00:02:33.292 00:02:33.292 Type BDF Vendor Device NUMA Driver Device Block devices 00:02:33.293 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:02:33.293 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme1 nvme1n1 00:02:33.293 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme0 nvme0n1 nvme0n2 nvme0n3 00:02:33.293 + rm -f /tmp/spdk-ld-path 00:02:33.293 + source autorun-spdk.conf 00:02:33.293 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:02:33.293 ++ SPDK_RUN_ASAN=1 00:02:33.293 ++ SPDK_RUN_UBSAN=1 00:02:33.293 ++ SPDK_TEST_RAID=1 00:02:33.293 ++ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:02:33.293 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:33.293 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:33.293 ++ RUN_NIGHTLY=1 00:02:33.293 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:02:33.293 + [[ -n '' ]] 00:02:33.293 + sudo git config --global --add safe.directory /home/vagrant/spdk_repo/spdk 00:02:33.293 + for M in /var/spdk/build-*-manifest.txt 00:02:33.293 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:02:33.293 + cp /var/spdk/build-kernel-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:33.572 + for M in /var/spdk/build-*-manifest.txt 00:02:33.572 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:02:33.572 + cp /var/spdk/build-pkg-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:33.572 + for M in /var/spdk/build-*-manifest.txt 00:02:33.572 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:02:33.572 + cp /var/spdk/build-repo-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:33.572 ++ uname 00:02:33.572 + [[ Linux == \L\i\n\u\x ]] 00:02:33.572 + sudo dmesg -T 00:02:33.572 + sudo dmesg --clear 00:02:33.572 + dmesg_pid=6157 00:02:33.572 + [[ Fedora Linux == FreeBSD ]] 00:02:33.572 + sudo dmesg -Tw 00:02:33.572 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:02:33.572 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:02:33.572 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:02:33.572 + [[ -x /usr/src/fio-static/fio ]] 00:02:33.572 + export FIO_BIN=/usr/src/fio-static/fio 00:02:33.572 + FIO_BIN=/usr/src/fio-static/fio 00:02:33.572 + [[ '' == \/\q\e\m\u\_\v\f\i\o\/* ]] 00:02:33.572 + [[ ! -v VFIO_QEMU_BIN ]] 00:02:33.572 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:02:33.572 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:02:33.572 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:02:33.572 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:02:33.572 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:02:33.572 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:02:33.572 + spdk/autorun.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:02:33.572 18:52:51 -- common/autotest_common.sh@1710 -- $ [[ n == y ]] 00:02:33.572 18:52:51 -- spdk/autorun.sh@20 -- $ source /home/vagrant/spdk_repo/autorun-spdk.conf 00:02:33.572 18:52:51 -- spdk_repo/autorun-spdk.conf@1 -- $ SPDK_RUN_FUNCTIONAL_TEST=1 00:02:33.572 18:52:51 -- spdk_repo/autorun-spdk.conf@2 -- $ SPDK_RUN_ASAN=1 00:02:33.572 18:52:51 -- spdk_repo/autorun-spdk.conf@3 -- $ SPDK_RUN_UBSAN=1 00:02:33.572 18:52:51 -- spdk_repo/autorun-spdk.conf@4 -- $ SPDK_TEST_RAID=1 00:02:33.572 18:52:51 -- spdk_repo/autorun-spdk.conf@5 -- $ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:02:33.572 18:52:51 -- spdk_repo/autorun-spdk.conf@6 -- $ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:33.572 18:52:51 -- spdk_repo/autorun-spdk.conf@7 -- $ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:33.573 18:52:51 -- spdk_repo/autorun-spdk.conf@8 -- $ RUN_NIGHTLY=1 00:02:33.573 18:52:51 -- spdk/autorun.sh@22 -- $ trap 'timing_finish || exit 1' EXIT 00:02:33.573 18:52:51 -- spdk/autorun.sh@25 -- $ /home/vagrant/spdk_repo/spdk/autobuild.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:02:33.832 18:52:51 -- common/autotest_common.sh@1710 -- $ [[ n == y ]] 00:02:33.832 18:52:51 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:02:33.832 18:52:51 -- scripts/common.sh@15 -- $ shopt -s extglob 00:02:33.832 18:52:51 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:02:33.833 18:52:51 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:02:33.833 18:52:51 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:02:33.833 18:52:51 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:33.833 18:52:51 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:33.833 18:52:51 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:33.833 18:52:51 -- paths/export.sh@5 -- $ export PATH 00:02:33.833 18:52:51 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:33.833 18:52:51 -- common/autobuild_common.sh@492 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:02:33.833 18:52:51 -- common/autobuild_common.sh@493 -- $ date +%s 00:02:33.833 18:52:51 -- common/autobuild_common.sh@493 -- $ mktemp -dt spdk_1733424771.XXXXXX 00:02:33.833 18:52:51 -- common/autobuild_common.sh@493 -- $ SPDK_WORKSPACE=/tmp/spdk_1733424771.U5DLwx 00:02:33.833 18:52:51 -- common/autobuild_common.sh@495 -- $ [[ -n '' ]] 00:02:33.833 18:52:51 -- common/autobuild_common.sh@499 -- $ '[' -n v22.11.4 ']' 00:02:33.833 18:52:51 -- common/autobuild_common.sh@500 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:02:33.833 18:52:51 -- common/autobuild_common.sh@500 -- $ scanbuild_exclude=' --exclude /home/vagrant/spdk_repo/dpdk' 00:02:33.833 18:52:51 -- common/autobuild_common.sh@506 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:02:33.833 18:52:51 -- common/autobuild_common.sh@508 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/dpdk --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:02:33.833 18:52:51 -- common/autobuild_common.sh@509 -- $ get_config_params 00:02:33.833 18:52:51 -- common/autotest_common.sh@409 -- $ xtrace_disable 00:02:33.833 18:52:51 -- common/autotest_common.sh@10 -- $ set +x 00:02:33.833 18:52:51 -- common/autobuild_common.sh@509 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build' 00:02:33.833 18:52:51 -- common/autobuild_common.sh@511 -- $ start_monitor_resources 00:02:33.833 18:52:51 -- pm/common@17 -- $ local monitor 00:02:33.833 18:52:51 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:33.833 18:52:51 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:33.833 18:52:51 -- pm/common@25 -- $ sleep 1 00:02:33.833 18:52:51 -- pm/common@21 -- $ date +%s 00:02:33.833 18:52:51 -- pm/common@21 -- $ date +%s 00:02:33.833 18:52:51 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1733424771 00:02:33.833 18:52:51 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1733424771 00:02:33.833 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1733424771_collect-vmstat.pm.log 00:02:33.833 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1733424771_collect-cpu-load.pm.log 00:02:34.774 18:52:52 -- common/autobuild_common.sh@512 -- $ trap stop_monitor_resources EXIT 00:02:34.774 18:52:52 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:02:34.774 18:52:52 -- spdk/autobuild.sh@12 -- $ umask 022 00:02:34.774 18:52:52 -- spdk/autobuild.sh@13 -- $ cd /home/vagrant/spdk_repo/spdk 00:02:34.774 18:52:52 -- spdk/autobuild.sh@16 -- $ date -u 00:02:34.774 Thu Dec 5 06:52:52 PM UTC 2024 00:02:34.774 18:52:52 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:02:34.774 v25.01-pre-296-g8d3947977 00:02:34.774 18:52:52 -- spdk/autobuild.sh@19 -- $ '[' 1 -eq 1 ']' 00:02:34.774 18:52:52 -- spdk/autobuild.sh@20 -- $ run_test asan echo 'using asan' 00:02:34.774 18:52:52 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:02:34.774 18:52:52 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:02:34.774 18:52:52 -- common/autotest_common.sh@10 -- $ set +x 00:02:34.774 ************************************ 00:02:34.774 START TEST asan 00:02:34.774 ************************************ 00:02:34.774 using asan 00:02:34.774 18:52:52 asan -- common/autotest_common.sh@1129 -- $ echo 'using asan' 00:02:34.774 00:02:34.774 real 0m0.001s 00:02:34.774 user 0m0.001s 00:02:34.774 sys 0m0.000s 00:02:34.774 18:52:52 asan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:02:34.774 18:52:52 asan -- common/autotest_common.sh@10 -- $ set +x 00:02:34.774 ************************************ 00:02:34.774 END TEST asan 00:02:34.774 ************************************ 00:02:35.170 18:52:52 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:02:35.170 18:52:52 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:02:35.170 18:52:52 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:02:35.170 18:52:52 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:02:35.170 18:52:52 -- common/autotest_common.sh@10 -- $ set +x 00:02:35.170 ************************************ 00:02:35.170 START TEST ubsan 00:02:35.170 ************************************ 00:02:35.170 using ubsan 00:02:35.170 18:52:52 ubsan -- common/autotest_common.sh@1129 -- $ echo 'using ubsan' 00:02:35.170 00:02:35.170 real 0m0.000s 00:02:35.170 user 0m0.000s 00:02:35.170 sys 0m0.000s 00:02:35.170 18:52:52 ubsan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:02:35.170 18:52:52 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:02:35.170 ************************************ 00:02:35.170 END TEST ubsan 00:02:35.170 ************************************ 00:02:35.170 18:52:52 -- spdk/autobuild.sh@27 -- $ '[' -n v22.11.4 ']' 00:02:35.170 18:52:52 -- spdk/autobuild.sh@28 -- $ build_native_dpdk 00:02:35.170 18:52:52 -- common/autobuild_common.sh@449 -- $ run_test build_native_dpdk _build_native_dpdk 00:02:35.170 18:52:52 -- common/autotest_common.sh@1105 -- $ '[' 2 -le 1 ']' 00:02:35.170 18:52:52 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:02:35.170 18:52:52 -- common/autotest_common.sh@10 -- $ set +x 00:02:35.170 ************************************ 00:02:35.170 START TEST build_native_dpdk 00:02:35.170 ************************************ 00:02:35.170 18:52:52 build_native_dpdk -- common/autotest_common.sh@1129 -- $ _build_native_dpdk 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@48 -- $ local external_dpdk_dir 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@49 -- $ local external_dpdk_base_dir 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@50 -- $ local compiler_version 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@51 -- $ local compiler 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@52 -- $ local dpdk_kmods 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@53 -- $ local repo=dpdk 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@55 -- $ compiler=gcc 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@61 -- $ export CC=gcc 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@61 -- $ CC=gcc 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *clang* ]] 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *gcc* ]] 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@68 -- $ gcc -dumpversion 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@68 -- $ compiler_version=13 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@69 -- $ compiler_version=13 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@70 -- $ external_dpdk_dir=/home/vagrant/spdk_repo/dpdk/build 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@71 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@71 -- $ external_dpdk_base_dir=/home/vagrant/spdk_repo/dpdk 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@73 -- $ [[ ! -d /home/vagrant/spdk_repo/dpdk ]] 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@82 -- $ orgdir=/home/vagrant/spdk_repo/spdk 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@83 -- $ git -C /home/vagrant/spdk_repo/dpdk log --oneline -n 5 00:02:35.170 caf0f5d395 version: 22.11.4 00:02:35.170 7d6f1cc05f Revert "net/iavf: fix abnormal disable HW interrupt" 00:02:35.170 dc9c799c7d vhost: fix missing spinlock unlock 00:02:35.170 4307659a90 net/mlx5: fix LACP redirection in Rx domain 00:02:35.170 6ef77f2a5e net/gve: fix RX buffer size alignment 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@85 -- $ dpdk_cflags='-fPIC -g -fcommon' 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@86 -- $ dpdk_ldflags= 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@87 -- $ dpdk_ver=22.11.4 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ gcc == *gcc* ]] 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ 13 -ge 5 ]] 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@90 -- $ dpdk_cflags+=' -Werror' 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ gcc == *gcc* ]] 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ 13 -ge 10 ]] 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@94 -- $ dpdk_cflags+=' -Wno-stringop-overflow' 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@102 -- $ DPDK_DRIVERS=("bus" "bus/pci" "bus/vdev" "mempool/ring" "net/i40e" "net/i40e/base" "power/acpi" "power/amd_pstate" "power/cppc" "power/intel_pstate" "power/intel_uncore" "power/kvm_vm") 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@103 -- $ local mlx5_libs_added=n 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@104 -- $ [[ 0 -eq 1 ]] 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@104 -- $ [[ 0 -eq 1 ]] 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@146 -- $ [[ 0 -eq 1 ]] 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@174 -- $ cd /home/vagrant/spdk_repo/dpdk 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@175 -- $ uname -s 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@175 -- $ '[' Linux = Linux ']' 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@176 -- $ lt 22.11.4 21.11.0 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 22.11.4 '<' 21.11.0 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 21 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@353 -- $ local d=21 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 21 =~ ^[0-9]+$ ]] 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@355 -- $ echo 21 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=21 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@367 -- $ return 1 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@180 -- $ patch -p1 00:02:35.170 patching file config/rte_config.h 00:02:35.170 Hunk #1 succeeded at 60 (offset 1 line). 00:02:35.170 18:52:52 build_native_dpdk -- common/autobuild_common.sh@183 -- $ lt 22.11.4 24.07.0 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 22.11.4 '<' 24.07.0 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:35.170 18:52:52 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@368 -- $ return 0 00:02:35.171 18:52:52 build_native_dpdk -- common/autobuild_common.sh@184 -- $ patch -p1 00:02:35.171 patching file lib/pcapng/rte_pcapng.c 00:02:35.171 Hunk #1 succeeded at 110 (offset -18 lines). 00:02:35.171 18:52:52 build_native_dpdk -- common/autobuild_common.sh@186 -- $ ge 22.11.4 24.07.0 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@376 -- $ cmp_versions 22.11.4 '>=' 24.07.0 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=>=' 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@348 -- $ : 1 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:02:35.171 18:52:52 build_native_dpdk -- scripts/common.sh@368 -- $ return 1 00:02:35.171 18:52:52 build_native_dpdk -- common/autobuild_common.sh@190 -- $ dpdk_kmods=false 00:02:35.171 18:52:52 build_native_dpdk -- common/autobuild_common.sh@191 -- $ uname -s 00:02:35.171 18:52:52 build_native_dpdk -- common/autobuild_common.sh@191 -- $ '[' Linux = FreeBSD ']' 00:02:35.171 18:52:52 build_native_dpdk -- common/autobuild_common.sh@195 -- $ printf %s, bus bus/pci bus/vdev mempool/ring net/i40e net/i40e/base power/acpi power/amd_pstate power/cppc power/intel_pstate power/intel_uncore power/kvm_vm 00:02:35.171 18:52:52 build_native_dpdk -- common/autobuild_common.sh@195 -- $ meson build-tmp --prefix=/home/vagrant/spdk_repo/dpdk/build --libdir lib -Denable_docs=false -Denable_kmods=false -Dtests=false -Dc_link_args= '-Dc_args=-fPIC -g -fcommon -Werror -Wno-stringop-overflow' -Dmachine=native -Denable_drivers=bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base,power/acpi,power/amd_pstate,power/cppc,power/intel_pstate,power/intel_uncore,power/kvm_vm, 00:02:41.742 The Meson build system 00:02:41.742 Version: 1.5.0 00:02:41.742 Source dir: /home/vagrant/spdk_repo/dpdk 00:02:41.742 Build dir: /home/vagrant/spdk_repo/dpdk/build-tmp 00:02:41.742 Build type: native build 00:02:41.742 Program cat found: YES (/usr/bin/cat) 00:02:41.742 Project name: DPDK 00:02:41.742 Project version: 22.11.4 00:02:41.742 C compiler for the host machine: gcc (gcc 13.3.1 "gcc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:02:41.742 C linker for the host machine: gcc ld.bfd 2.40-14 00:02:41.742 Host machine cpu family: x86_64 00:02:41.742 Host machine cpu: x86_64 00:02:41.742 Message: ## Building in Developer Mode ## 00:02:41.742 Program pkg-config found: YES (/usr/bin/pkg-config) 00:02:41.742 Program check-symbols.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/check-symbols.sh) 00:02:41.742 Program options-ibverbs-static.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/options-ibverbs-static.sh) 00:02:41.742 Program objdump found: YES (/usr/bin/objdump) 00:02:41.742 Program python3 found: YES (/usr/bin/python3) 00:02:41.742 Program cat found: YES (/usr/bin/cat) 00:02:41.742 config/meson.build:83: WARNING: The "machine" option is deprecated. Please use "cpu_instruction_set" instead. 00:02:41.742 Checking for size of "void *" : 8 00:02:41.742 Checking for size of "void *" : 8 (cached) 00:02:41.742 Library m found: YES 00:02:41.742 Library numa found: YES 00:02:41.742 Has header "numaif.h" : YES 00:02:41.742 Library fdt found: NO 00:02:41.742 Library execinfo found: NO 00:02:41.742 Has header "execinfo.h" : YES 00:02:41.742 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:02:41.742 Run-time dependency libarchive found: NO (tried pkgconfig) 00:02:41.742 Run-time dependency libbsd found: NO (tried pkgconfig) 00:02:41.742 Run-time dependency jansson found: NO (tried pkgconfig) 00:02:41.742 Run-time dependency openssl found: YES 3.1.1 00:02:41.742 Run-time dependency libpcap found: YES 1.10.4 00:02:41.742 Has header "pcap.h" with dependency libpcap: YES 00:02:41.742 Compiler for C supports arguments -Wcast-qual: YES 00:02:41.742 Compiler for C supports arguments -Wdeprecated: YES 00:02:41.742 Compiler for C supports arguments -Wformat: YES 00:02:41.742 Compiler for C supports arguments -Wformat-nonliteral: NO 00:02:41.742 Compiler for C supports arguments -Wformat-security: NO 00:02:41.742 Compiler for C supports arguments -Wmissing-declarations: YES 00:02:41.742 Compiler for C supports arguments -Wmissing-prototypes: YES 00:02:41.742 Compiler for C supports arguments -Wnested-externs: YES 00:02:41.742 Compiler for C supports arguments -Wold-style-definition: YES 00:02:41.742 Compiler for C supports arguments -Wpointer-arith: YES 00:02:41.742 Compiler for C supports arguments -Wsign-compare: YES 00:02:41.742 Compiler for C supports arguments -Wstrict-prototypes: YES 00:02:41.742 Compiler for C supports arguments -Wundef: YES 00:02:41.742 Compiler for C supports arguments -Wwrite-strings: YES 00:02:41.742 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:02:41.742 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:02:41.742 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:02:41.742 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:02:41.742 Compiler for C supports arguments -mavx512f: YES 00:02:41.742 Checking if "AVX512 checking" compiles: YES 00:02:41.742 Fetching value of define "__SSE4_2__" : 1 00:02:41.742 Fetching value of define "__AES__" : 1 00:02:41.742 Fetching value of define "__AVX__" : 1 00:02:41.742 Fetching value of define "__AVX2__" : 1 00:02:41.742 Fetching value of define "__AVX512BW__" : 1 00:02:41.742 Fetching value of define "__AVX512CD__" : 1 00:02:41.742 Fetching value of define "__AVX512DQ__" : 1 00:02:41.742 Fetching value of define "__AVX512F__" : 1 00:02:41.742 Fetching value of define "__AVX512VL__" : 1 00:02:41.742 Fetching value of define "__PCLMUL__" : 1 00:02:41.742 Fetching value of define "__RDRND__" : 1 00:02:41.742 Fetching value of define "__RDSEED__" : 1 00:02:41.742 Fetching value of define "__VPCLMULQDQ__" : (undefined) 00:02:41.742 Compiler for C supports arguments -Wno-format-truncation: YES 00:02:41.742 Message: lib/kvargs: Defining dependency "kvargs" 00:02:41.742 Message: lib/telemetry: Defining dependency "telemetry" 00:02:41.742 Checking for function "getentropy" : YES 00:02:41.742 Message: lib/eal: Defining dependency "eal" 00:02:41.742 Message: lib/ring: Defining dependency "ring" 00:02:41.742 Message: lib/rcu: Defining dependency "rcu" 00:02:41.742 Message: lib/mempool: Defining dependency "mempool" 00:02:41.742 Message: lib/mbuf: Defining dependency "mbuf" 00:02:41.742 Fetching value of define "__PCLMUL__" : 1 (cached) 00:02:41.742 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:41.742 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:41.742 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:41.742 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:41.742 Fetching value of define "__VPCLMULQDQ__" : (undefined) (cached) 00:02:41.742 Compiler for C supports arguments -mpclmul: YES 00:02:41.742 Compiler for C supports arguments -maes: YES 00:02:41.742 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:41.742 Compiler for C supports arguments -mavx512bw: YES 00:02:41.742 Compiler for C supports arguments -mavx512dq: YES 00:02:41.742 Compiler for C supports arguments -mavx512vl: YES 00:02:41.742 Compiler for C supports arguments -mvpclmulqdq: YES 00:02:41.742 Compiler for C supports arguments -mavx2: YES 00:02:41.742 Compiler for C supports arguments -mavx: YES 00:02:41.742 Message: lib/net: Defining dependency "net" 00:02:41.742 Message: lib/meter: Defining dependency "meter" 00:02:41.742 Message: lib/ethdev: Defining dependency "ethdev" 00:02:41.742 Message: lib/pci: Defining dependency "pci" 00:02:41.742 Message: lib/cmdline: Defining dependency "cmdline" 00:02:41.742 Message: lib/metrics: Defining dependency "metrics" 00:02:41.742 Message: lib/hash: Defining dependency "hash" 00:02:41.742 Message: lib/timer: Defining dependency "timer" 00:02:41.742 Fetching value of define "__AVX2__" : 1 (cached) 00:02:41.742 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:41.742 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:41.742 Fetching value of define "__AVX512CD__" : 1 (cached) 00:02:41.742 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:41.742 Message: lib/acl: Defining dependency "acl" 00:02:41.742 Message: lib/bbdev: Defining dependency "bbdev" 00:02:41.742 Message: lib/bitratestats: Defining dependency "bitratestats" 00:02:41.742 Run-time dependency libelf found: YES 0.191 00:02:41.742 Message: lib/bpf: Defining dependency "bpf" 00:02:41.742 Message: lib/cfgfile: Defining dependency "cfgfile" 00:02:41.742 Message: lib/compressdev: Defining dependency "compressdev" 00:02:41.742 Message: lib/cryptodev: Defining dependency "cryptodev" 00:02:41.742 Message: lib/distributor: Defining dependency "distributor" 00:02:41.742 Message: lib/efd: Defining dependency "efd" 00:02:41.742 Message: lib/eventdev: Defining dependency "eventdev" 00:02:41.742 Message: lib/gpudev: Defining dependency "gpudev" 00:02:41.742 Message: lib/gro: Defining dependency "gro" 00:02:41.742 Message: lib/gso: Defining dependency "gso" 00:02:41.742 Message: lib/ip_frag: Defining dependency "ip_frag" 00:02:41.742 Message: lib/jobstats: Defining dependency "jobstats" 00:02:41.742 Message: lib/latencystats: Defining dependency "latencystats" 00:02:41.742 Message: lib/lpm: Defining dependency "lpm" 00:02:41.742 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:41.742 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:41.742 Fetching value of define "__AVX512IFMA__" : (undefined) 00:02:41.742 Compiler for C supports arguments -mavx512f -mavx512dq -mavx512ifma: YES 00:02:41.742 Message: lib/member: Defining dependency "member" 00:02:41.742 Message: lib/pcapng: Defining dependency "pcapng" 00:02:41.742 Compiler for C supports arguments -Wno-cast-qual: YES 00:02:41.742 Message: lib/power: Defining dependency "power" 00:02:41.742 Message: lib/rawdev: Defining dependency "rawdev" 00:02:41.742 Message: lib/regexdev: Defining dependency "regexdev" 00:02:41.742 Message: lib/dmadev: Defining dependency "dmadev" 00:02:41.743 Message: lib/rib: Defining dependency "rib" 00:02:41.743 Message: lib/reorder: Defining dependency "reorder" 00:02:41.743 Message: lib/sched: Defining dependency "sched" 00:02:41.743 Message: lib/security: Defining dependency "security" 00:02:41.743 Message: lib/stack: Defining dependency "stack" 00:02:41.743 Has header "linux/userfaultfd.h" : YES 00:02:41.743 Message: lib/vhost: Defining dependency "vhost" 00:02:41.743 Message: lib/ipsec: Defining dependency "ipsec" 00:02:41.743 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:41.743 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:41.743 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:41.743 Message: lib/fib: Defining dependency "fib" 00:02:41.743 Message: lib/port: Defining dependency "port" 00:02:41.743 Message: lib/pdump: Defining dependency "pdump" 00:02:41.743 Message: lib/table: Defining dependency "table" 00:02:41.743 Message: lib/pipeline: Defining dependency "pipeline" 00:02:41.743 Message: lib/graph: Defining dependency "graph" 00:02:41.743 Message: lib/node: Defining dependency "node" 00:02:41.743 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:02:41.743 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:02:41.743 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:02:41.743 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:02:41.743 Compiler for C supports arguments -Wno-sign-compare: YES 00:02:41.743 Compiler for C supports arguments -Wno-unused-value: YES 00:02:41.743 Compiler for C supports arguments -Wno-format: YES 00:02:41.743 Compiler for C supports arguments -Wno-format-security: YES 00:02:41.743 Compiler for C supports arguments -Wno-format-nonliteral: YES 00:02:41.743 Compiler for C supports arguments -Wno-strict-aliasing: YES 00:02:42.001 Compiler for C supports arguments -Wno-unused-but-set-variable: YES 00:02:42.001 Compiler for C supports arguments -Wno-unused-parameter: YES 00:02:42.001 Fetching value of define "__AVX2__" : 1 (cached) 00:02:42.001 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:42.001 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:42.001 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:42.001 Compiler for C supports arguments -mavx512bw: YES (cached) 00:02:42.001 Compiler for C supports arguments -march=skylake-avx512: YES 00:02:42.001 Message: drivers/net/i40e: Defining dependency "net_i40e" 00:02:42.001 Program doxygen found: YES (/usr/local/bin/doxygen) 00:02:42.001 Configuring doxy-api.conf using configuration 00:02:42.001 Program sphinx-build found: NO 00:02:42.001 Configuring rte_build_config.h using configuration 00:02:42.001 Message: 00:02:42.001 ================= 00:02:42.001 Applications Enabled 00:02:42.001 ================= 00:02:42.001 00:02:42.001 apps: 00:02:42.001 dumpcap, pdump, proc-info, test-acl, test-bbdev, test-cmdline, test-compress-perf, test-crypto-perf, 00:02:42.001 test-eventdev, test-fib, test-flow-perf, test-gpudev, test-pipeline, test-pmd, test-regex, test-sad, 00:02:42.001 test-security-perf, 00:02:42.001 00:02:42.001 Message: 00:02:42.001 ================= 00:02:42.001 Libraries Enabled 00:02:42.001 ================= 00:02:42.001 00:02:42.001 libs: 00:02:42.001 kvargs, telemetry, eal, ring, rcu, mempool, mbuf, net, 00:02:42.001 meter, ethdev, pci, cmdline, metrics, hash, timer, acl, 00:02:42.001 bbdev, bitratestats, bpf, cfgfile, compressdev, cryptodev, distributor, efd, 00:02:42.001 eventdev, gpudev, gro, gso, ip_frag, jobstats, latencystats, lpm, 00:02:42.001 member, pcapng, power, rawdev, regexdev, dmadev, rib, reorder, 00:02:42.001 sched, security, stack, vhost, ipsec, fib, port, pdump, 00:02:42.001 table, pipeline, graph, node, 00:02:42.001 00:02:42.001 Message: 00:02:42.001 =============== 00:02:42.001 Drivers Enabled 00:02:42.001 =============== 00:02:42.001 00:02:42.001 common: 00:02:42.001 00:02:42.001 bus: 00:02:42.001 pci, vdev, 00:02:42.001 mempool: 00:02:42.001 ring, 00:02:42.001 dma: 00:02:42.001 00:02:42.001 net: 00:02:42.001 i40e, 00:02:42.001 raw: 00:02:42.001 00:02:42.001 crypto: 00:02:42.001 00:02:42.001 compress: 00:02:42.001 00:02:42.001 regex: 00:02:42.001 00:02:42.001 vdpa: 00:02:42.001 00:02:42.001 event: 00:02:42.001 00:02:42.001 baseband: 00:02:42.001 00:02:42.001 gpu: 00:02:42.001 00:02:42.001 00:02:42.001 Message: 00:02:42.001 ================= 00:02:42.001 Content Skipped 00:02:42.001 ================= 00:02:42.001 00:02:42.001 apps: 00:02:42.001 00:02:42.001 libs: 00:02:42.001 kni: explicitly disabled via build config (deprecated lib) 00:02:42.001 flow_classify: explicitly disabled via build config (deprecated lib) 00:02:42.001 00:02:42.001 drivers: 00:02:42.001 common/cpt: not in enabled drivers build config 00:02:42.001 common/dpaax: not in enabled drivers build config 00:02:42.001 common/iavf: not in enabled drivers build config 00:02:42.001 common/idpf: not in enabled drivers build config 00:02:42.001 common/mvep: not in enabled drivers build config 00:02:42.001 common/octeontx: not in enabled drivers build config 00:02:42.001 bus/auxiliary: not in enabled drivers build config 00:02:42.001 bus/dpaa: not in enabled drivers build config 00:02:42.001 bus/fslmc: not in enabled drivers build config 00:02:42.001 bus/ifpga: not in enabled drivers build config 00:02:42.001 bus/vmbus: not in enabled drivers build config 00:02:42.001 common/cnxk: not in enabled drivers build config 00:02:42.001 common/mlx5: not in enabled drivers build config 00:02:42.001 common/qat: not in enabled drivers build config 00:02:42.001 common/sfc_efx: not in enabled drivers build config 00:02:42.001 mempool/bucket: not in enabled drivers build config 00:02:42.001 mempool/cnxk: not in enabled drivers build config 00:02:42.001 mempool/dpaa: not in enabled drivers build config 00:02:42.001 mempool/dpaa2: not in enabled drivers build config 00:02:42.001 mempool/octeontx: not in enabled drivers build config 00:02:42.001 mempool/stack: not in enabled drivers build config 00:02:42.001 dma/cnxk: not in enabled drivers build config 00:02:42.001 dma/dpaa: not in enabled drivers build config 00:02:42.001 dma/dpaa2: not in enabled drivers build config 00:02:42.001 dma/hisilicon: not in enabled drivers build config 00:02:42.001 dma/idxd: not in enabled drivers build config 00:02:42.001 dma/ioat: not in enabled drivers build config 00:02:42.002 dma/skeleton: not in enabled drivers build config 00:02:42.002 net/af_packet: not in enabled drivers build config 00:02:42.002 net/af_xdp: not in enabled drivers build config 00:02:42.002 net/ark: not in enabled drivers build config 00:02:42.002 net/atlantic: not in enabled drivers build config 00:02:42.002 net/avp: not in enabled drivers build config 00:02:42.002 net/axgbe: not in enabled drivers build config 00:02:42.002 net/bnx2x: not in enabled drivers build config 00:02:42.002 net/bnxt: not in enabled drivers build config 00:02:42.002 net/bonding: not in enabled drivers build config 00:02:42.002 net/cnxk: not in enabled drivers build config 00:02:42.002 net/cxgbe: not in enabled drivers build config 00:02:42.002 net/dpaa: not in enabled drivers build config 00:02:42.002 net/dpaa2: not in enabled drivers build config 00:02:42.002 net/e1000: not in enabled drivers build config 00:02:42.002 net/ena: not in enabled drivers build config 00:02:42.002 net/enetc: not in enabled drivers build config 00:02:42.002 net/enetfec: not in enabled drivers build config 00:02:42.002 net/enic: not in enabled drivers build config 00:02:42.002 net/failsafe: not in enabled drivers build config 00:02:42.002 net/fm10k: not in enabled drivers build config 00:02:42.002 net/gve: not in enabled drivers build config 00:02:42.002 net/hinic: not in enabled drivers build config 00:02:42.002 net/hns3: not in enabled drivers build config 00:02:42.002 net/iavf: not in enabled drivers build config 00:02:42.002 net/ice: not in enabled drivers build config 00:02:42.002 net/idpf: not in enabled drivers build config 00:02:42.002 net/igc: not in enabled drivers build config 00:02:42.002 net/ionic: not in enabled drivers build config 00:02:42.002 net/ipn3ke: not in enabled drivers build config 00:02:42.002 net/ixgbe: not in enabled drivers build config 00:02:42.002 net/kni: not in enabled drivers build config 00:02:42.002 net/liquidio: not in enabled drivers build config 00:02:42.002 net/mana: not in enabled drivers build config 00:02:42.002 net/memif: not in enabled drivers build config 00:02:42.002 net/mlx4: not in enabled drivers build config 00:02:42.002 net/mlx5: not in enabled drivers build config 00:02:42.002 net/mvneta: not in enabled drivers build config 00:02:42.002 net/mvpp2: not in enabled drivers build config 00:02:42.002 net/netvsc: not in enabled drivers build config 00:02:42.002 net/nfb: not in enabled drivers build config 00:02:42.002 net/nfp: not in enabled drivers build config 00:02:42.002 net/ngbe: not in enabled drivers build config 00:02:42.002 net/null: not in enabled drivers build config 00:02:42.002 net/octeontx: not in enabled drivers build config 00:02:42.002 net/octeon_ep: not in enabled drivers build config 00:02:42.002 net/pcap: not in enabled drivers build config 00:02:42.002 net/pfe: not in enabled drivers build config 00:02:42.002 net/qede: not in enabled drivers build config 00:02:42.002 net/ring: not in enabled drivers build config 00:02:42.002 net/sfc: not in enabled drivers build config 00:02:42.002 net/softnic: not in enabled drivers build config 00:02:42.002 net/tap: not in enabled drivers build config 00:02:42.002 net/thunderx: not in enabled drivers build config 00:02:42.002 net/txgbe: not in enabled drivers build config 00:02:42.002 net/vdev_netvsc: not in enabled drivers build config 00:02:42.002 net/vhost: not in enabled drivers build config 00:02:42.002 net/virtio: not in enabled drivers build config 00:02:42.002 net/vmxnet3: not in enabled drivers build config 00:02:42.002 raw/cnxk_bphy: not in enabled drivers build config 00:02:42.002 raw/cnxk_gpio: not in enabled drivers build config 00:02:42.002 raw/dpaa2_cmdif: not in enabled drivers build config 00:02:42.002 raw/ifpga: not in enabled drivers build config 00:02:42.002 raw/ntb: not in enabled drivers build config 00:02:42.002 raw/skeleton: not in enabled drivers build config 00:02:42.002 crypto/armv8: not in enabled drivers build config 00:02:42.002 crypto/bcmfs: not in enabled drivers build config 00:02:42.002 crypto/caam_jr: not in enabled drivers build config 00:02:42.002 crypto/ccp: not in enabled drivers build config 00:02:42.002 crypto/cnxk: not in enabled drivers build config 00:02:42.002 crypto/dpaa_sec: not in enabled drivers build config 00:02:42.002 crypto/dpaa2_sec: not in enabled drivers build config 00:02:42.002 crypto/ipsec_mb: not in enabled drivers build config 00:02:42.002 crypto/mlx5: not in enabled drivers build config 00:02:42.002 crypto/mvsam: not in enabled drivers build config 00:02:42.002 crypto/nitrox: not in enabled drivers build config 00:02:42.002 crypto/null: not in enabled drivers build config 00:02:42.002 crypto/octeontx: not in enabled drivers build config 00:02:42.002 crypto/openssl: not in enabled drivers build config 00:02:42.002 crypto/scheduler: not in enabled drivers build config 00:02:42.002 crypto/uadk: not in enabled drivers build config 00:02:42.002 crypto/virtio: not in enabled drivers build config 00:02:42.002 compress/isal: not in enabled drivers build config 00:02:42.002 compress/mlx5: not in enabled drivers build config 00:02:42.002 compress/octeontx: not in enabled drivers build config 00:02:42.002 compress/zlib: not in enabled drivers build config 00:02:42.002 regex/mlx5: not in enabled drivers build config 00:02:42.002 regex/cn9k: not in enabled drivers build config 00:02:42.002 vdpa/ifc: not in enabled drivers build config 00:02:42.002 vdpa/mlx5: not in enabled drivers build config 00:02:42.002 vdpa/sfc: not in enabled drivers build config 00:02:42.002 event/cnxk: not in enabled drivers build config 00:02:42.002 event/dlb2: not in enabled drivers build config 00:02:42.002 event/dpaa: not in enabled drivers build config 00:02:42.002 event/dpaa2: not in enabled drivers build config 00:02:42.002 event/dsw: not in enabled drivers build config 00:02:42.002 event/opdl: not in enabled drivers build config 00:02:42.002 event/skeleton: not in enabled drivers build config 00:02:42.002 event/sw: not in enabled drivers build config 00:02:42.002 event/octeontx: not in enabled drivers build config 00:02:42.002 baseband/acc: not in enabled drivers build config 00:02:42.002 baseband/fpga_5gnr_fec: not in enabled drivers build config 00:02:42.002 baseband/fpga_lte_fec: not in enabled drivers build config 00:02:42.002 baseband/la12xx: not in enabled drivers build config 00:02:42.002 baseband/null: not in enabled drivers build config 00:02:42.002 baseband/turbo_sw: not in enabled drivers build config 00:02:42.002 gpu/cuda: not in enabled drivers build config 00:02:42.002 00:02:42.002 00:02:42.002 Build targets in project: 311 00:02:42.002 00:02:42.002 DPDK 22.11.4 00:02:42.002 00:02:42.002 User defined options 00:02:42.002 libdir : lib 00:02:42.002 prefix : /home/vagrant/spdk_repo/dpdk/build 00:02:42.002 c_args : -fPIC -g -fcommon -Werror -Wno-stringop-overflow 00:02:42.002 c_link_args : 00:02:42.002 enable_docs : false 00:02:42.002 enable_drivers: bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base,power/acpi,power/amd_pstate,power/cppc,power/intel_pstate,power/intel_uncore,power/kvm_vm, 00:02:42.002 enable_kmods : false 00:02:42.002 machine : native 00:02:42.002 tests : false 00:02:42.002 00:02:42.002 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:02:42.002 WARNING: Running the setup command as `meson [options]` instead of `meson setup [options]` is ambiguous and deprecated. 00:02:42.002 18:52:59 build_native_dpdk -- common/autobuild_common.sh@199 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 00:02:42.002 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:02:42.002 [1/740] Generating lib/rte_kvargs_mingw with a custom command 00:02:42.002 [2/740] Generating lib/rte_telemetry_mingw with a custom command 00:02:42.002 [3/740] Generating lib/rte_telemetry_def with a custom command 00:02:42.002 [4/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:02:42.002 [5/740] Generating lib/rte_kvargs_def with a custom command 00:02:42.261 [6/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:02:42.261 [7/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:02:42.261 [8/740] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:02:42.261 [9/740] Linking static target lib/librte_kvargs.a 00:02:42.261 [10/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:02:42.261 [11/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:02:42.261 [12/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:02:42.261 [13/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:02:42.261 [14/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:02:42.261 [15/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:02:42.261 [16/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:02:42.261 [17/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:02:42.261 [18/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:02:42.261 [19/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:02:42.261 [20/740] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:02:42.261 [21/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:02:42.519 [22/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_log.c.o 00:02:42.519 [23/740] Linking target lib/librte_kvargs.so.23.0 00:02:42.519 [24/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:02:42.519 [25/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:02:42.519 [26/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:02:42.519 [27/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:02:42.519 [28/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:02:42.519 [29/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:02:42.519 [30/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:02:42.519 [31/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:02:42.519 [32/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:02:42.519 [33/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:02:42.519 [34/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:02:42.519 [35/740] Linking static target lib/librte_telemetry.a 00:02:42.519 [36/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:02:42.777 [37/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:02:42.777 [38/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:02:42.777 [39/740] Generating symbol file lib/librte_kvargs.so.23.0.p/librte_kvargs.so.23.0.symbols 00:02:42.777 [40/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:02:42.777 [41/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:02:42.777 [42/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:02:42.777 [43/740] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:02:42.777 [44/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:02:42.777 [45/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:02:43.036 [46/740] Linking target lib/librte_telemetry.so.23.0 00:02:43.037 [47/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:02:43.037 [48/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:02:43.037 [49/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:02:43.037 [50/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:02:43.037 [51/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:02:43.037 [52/740] Generating symbol file lib/librte_telemetry.so.23.0.p/librte_telemetry.so.23.0.symbols 00:02:43.037 [53/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:02:43.037 [54/740] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:02:43.037 [55/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:02:43.037 [56/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:02:43.037 [57/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:02:43.037 [58/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:02:43.037 [59/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:02:43.037 [60/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:02:43.037 [61/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:02:43.037 [62/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:02:43.037 [63/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:02:43.037 [64/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:02:43.037 [65/740] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:02:43.037 [66/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_log.c.o 00:02:43.295 [67/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:02:43.295 [68/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:02:43.295 [69/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:02:43.295 [70/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:02:43.295 [71/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:02:43.295 [72/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:02:43.295 [73/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:02:43.295 [74/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:02:43.295 [75/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:02:43.295 [76/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:02:43.295 [77/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:02:43.295 [78/740] Generating lib/rte_eal_def with a custom command 00:02:43.295 [79/740] Generating lib/rte_eal_mingw with a custom command 00:02:43.295 [80/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:02:43.295 [81/740] Generating lib/rte_ring_def with a custom command 00:02:43.295 [82/740] Generating lib/rte_ring_mingw with a custom command 00:02:43.295 [83/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:02:43.295 [84/740] Generating lib/rte_rcu_def with a custom command 00:02:43.295 [85/740] Generating lib/rte_rcu_mingw with a custom command 00:02:43.295 [86/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:02:43.553 [87/740] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:02:43.553 [88/740] Linking static target lib/librte_ring.a 00:02:43.553 [89/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:02:43.553 [90/740] Generating lib/rte_mempool_def with a custom command 00:02:43.553 [91/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:02:43.553 [92/740] Generating lib/rte_mempool_mingw with a custom command 00:02:43.553 [93/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:02:43.553 [94/740] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:02:43.812 [95/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:02:43.812 [96/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:02:43.812 [97/740] Linking static target lib/librte_eal.a 00:02:43.812 [98/740] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:02:43.812 [99/740] Generating lib/rte_mbuf_def with a custom command 00:02:43.812 [100/740] Generating lib/rte_mbuf_mingw with a custom command 00:02:43.812 [101/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:02:43.812 [102/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:02:43.812 [103/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:02:44.069 [104/740] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:02:44.069 [105/740] Linking static target lib/librte_rcu.a 00:02:44.069 [106/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:02:44.069 [107/740] Linking static target lib/librte_mempool.a 00:02:44.069 [108/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:02:44.327 [109/740] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:02:44.327 [110/740] Generating lib/rte_net_def with a custom command 00:02:44.327 [111/740] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:02:44.327 [112/740] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:02:44.327 [113/740] Compiling C object lib/net/libnet_crc_avx512_lib.a.p/net_crc_avx512.c.o 00:02:44.327 [114/740] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:02:44.327 [115/740] Generating lib/rte_net_mingw with a custom command 00:02:44.327 [116/740] Linking static target lib/net/libnet_crc_avx512_lib.a 00:02:44.327 [117/740] Generating lib/rte_meter_mingw with a custom command 00:02:44.327 [118/740] Generating lib/rte_meter_def with a custom command 00:02:44.327 [119/740] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:02:44.327 [120/740] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:02:44.327 [121/740] Linking static target lib/librte_meter.a 00:02:44.584 [122/740] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:02:44.584 [123/740] Linking static target lib/librte_net.a 00:02:44.584 [124/740] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:02:44.584 [125/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:02:44.584 [126/740] Linking static target lib/librte_mbuf.a 00:02:44.584 [127/740] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:02:44.584 [128/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:02:44.584 [129/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:02:44.852 [130/740] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:02:44.852 [131/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:02:44.852 [132/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:02:44.852 [133/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:02:45.110 [134/740] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:45.110 [135/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:02:45.110 [136/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:02:45.110 [137/740] Generating lib/rte_ethdev_def with a custom command 00:02:45.110 [138/740] Generating lib/rte_ethdev_mingw with a custom command 00:02:45.110 [139/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:02:45.110 [140/740] Generating lib/rte_pci_def with a custom command 00:02:45.368 [141/740] Generating lib/rte_pci_mingw with a custom command 00:02:45.368 [142/740] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:02:45.368 [143/740] Linking static target lib/librte_pci.a 00:02:45.368 [144/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:02:45.368 [145/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:02:45.368 [146/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:02:45.368 [147/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:02:45.368 [148/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:02:45.368 [149/740] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:45.368 [150/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:02:45.368 [151/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:02:45.368 [152/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:02:45.627 [153/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:02:45.627 [154/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:02:45.627 [155/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:02:45.627 [156/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:02:45.627 [157/740] Generating lib/rte_cmdline_def with a custom command 00:02:45.627 [158/740] Generating lib/rte_cmdline_mingw with a custom command 00:02:45.627 [159/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:02:45.627 [160/740] Generating lib/rte_metrics_def with a custom command 00:02:45.627 [161/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:02:45.627 [162/740] Generating lib/rte_metrics_mingw with a custom command 00:02:45.627 [163/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:02:45.627 [164/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:02:45.627 [165/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:02:45.627 [166/740] Linking static target lib/librte_cmdline.a 00:02:45.627 [167/740] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics.c.o 00:02:45.627 [168/740] Generating lib/rte_hash_def with a custom command 00:02:45.627 [169/740] Generating lib/rte_hash_mingw with a custom command 00:02:45.627 [170/740] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:02:45.885 [171/740] Generating lib/rte_timer_def with a custom command 00:02:45.885 [172/740] Generating lib/rte_timer_mingw with a custom command 00:02:45.885 [173/740] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics_telemetry.c.o 00:02:45.885 [174/740] Linking static target lib/librte_metrics.a 00:02:46.144 [175/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:02:46.144 [176/740] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:02:46.144 [177/740] Linking static target lib/librte_timer.a 00:02:46.144 [178/740] Generating lib/metrics.sym_chk with a custom command (wrapped by meson to capture output) 00:02:46.403 [179/740] Compiling C object lib/librte_acl.a.p/acl_acl_gen.c.o 00:02:46.403 [180/740] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:02:46.403 [181/740] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:02:46.403 [182/740] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:02:46.662 [183/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_scalar.c.o 00:02:46.662 [184/740] Generating lib/rte_acl_def with a custom command 00:02:46.662 [185/740] Compiling C object lib/librte_acl.a.p/acl_tb_mem.c.o 00:02:46.662 [186/740] Generating lib/rte_acl_mingw with a custom command 00:02:46.662 [187/740] Generating lib/rte_bbdev_def with a custom command 00:02:46.662 [188/740] Compiling C object lib/librte_acl.a.p/acl_rte_acl.c.o 00:02:46.662 [189/740] Generating lib/rte_bbdev_mingw with a custom command 00:02:46.662 [190/740] Generating lib/rte_bitratestats_def with a custom command 00:02:46.662 [191/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:02:46.662 [192/740] Generating lib/rte_bitratestats_mingw with a custom command 00:02:46.662 [193/740] Linking static target lib/librte_ethdev.a 00:02:47.230 [194/740] Compiling C object lib/librte_bitratestats.a.p/bitratestats_rte_bitrate.c.o 00:02:47.230 [195/740] Linking static target lib/librte_bitratestats.a 00:02:47.230 [196/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf.c.o 00:02:47.230 [197/740] Compiling C object lib/librte_acl.a.p/acl_acl_bld.c.o 00:02:47.230 [198/740] Compiling C object lib/librte_bbdev.a.p/bbdev_rte_bbdev.c.o 00:02:47.230 [199/740] Linking static target lib/librte_bbdev.a 00:02:47.230 [200/740] Generating lib/bitratestats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.489 [201/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_dump.c.o 00:02:47.747 [202/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load.c.o 00:02:47.747 [203/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_exec.c.o 00:02:47.747 [204/740] Generating lib/bbdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.005 [205/740] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:02:48.005 [206/740] Linking static target lib/librte_hash.a 00:02:48.005 [207/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_sse.c.o 00:02:48.005 [208/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_stub.c.o 00:02:48.264 [209/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_pkt.c.o 00:02:48.264 [210/740] Generating lib/rte_bpf_def with a custom command 00:02:48.264 [211/740] Generating lib/rte_bpf_mingw with a custom command 00:02:48.523 [212/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load_elf.c.o 00:02:48.523 [213/740] Generating lib/rte_cfgfile_def with a custom command 00:02:48.523 [214/740] Generating lib/rte_cfgfile_mingw with a custom command 00:02:48.523 [215/740] Compiling C object lib/librte_cfgfile.a.p/cfgfile_rte_cfgfile.c.o 00:02:48.523 [216/740] Linking static target lib/librte_cfgfile.a 00:02:48.523 [217/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_validate.c.o 00:02:48.523 [218/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_convert.c.o 00:02:48.523 [219/740] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.523 [220/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_avx2.c.o 00:02:48.523 [221/740] Generating lib/rte_compressdev_def with a custom command 00:02:48.523 [222/740] Generating lib/rte_compressdev_mingw with a custom command 00:02:48.783 [223/740] Generating lib/cfgfile.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.783 [224/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_jit_x86.c.o 00:02:48.783 [225/740] Linking static target lib/librte_bpf.a 00:02:48.783 [226/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:02:48.783 [227/740] Generating lib/rte_cryptodev_def with a custom command 00:02:48.783 [228/740] Generating lib/rte_cryptodev_mingw with a custom command 00:02:49.043 [229/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:02:49.043 [230/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:02:49.043 [231/740] Linking static target lib/librte_compressdev.a 00:02:49.043 [232/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:02:49.043 [233/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_avx512.c.o 00:02:49.043 [234/740] Generating lib/bpf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:49.043 [235/740] Linking static target lib/librte_acl.a 00:02:49.043 [236/740] Generating lib/rte_distributor_def with a custom command 00:02:49.043 [237/740] Generating lib/rte_distributor_mingw with a custom command 00:02:49.043 [238/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:02:49.043 [239/740] Generating lib/rte_efd_def with a custom command 00:02:49.302 [240/740] Generating lib/rte_efd_mingw with a custom command 00:02:49.302 [241/740] Generating lib/acl.sym_chk with a custom command (wrapped by meson to capture output) 00:02:49.302 [242/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_match_sse.c.o 00:02:49.302 [243/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_single.c.o 00:02:49.561 [244/740] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:02:49.561 [245/740] Linking target lib/librte_eal.so.23.0 00:02:49.561 [246/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor.c.o 00:02:49.561 [247/740] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_private.c.o 00:02:49.561 [248/740] Linking static target lib/librte_distributor.a 00:02:49.561 [249/740] Generating symbol file lib/librte_eal.so.23.0.p/librte_eal.so.23.0.symbols 00:02:49.561 [250/740] Linking target lib/librte_ring.so.23.0 00:02:49.820 [251/740] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:49.820 [252/740] Linking target lib/librte_meter.so.23.0 00:02:49.820 [253/740] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_trace_points.c.o 00:02:49.820 [254/740] Generating symbol file lib/librte_ring.so.23.0.p/librte_ring.so.23.0.symbols 00:02:49.820 [255/740] Linking target lib/librte_rcu.so.23.0 00:02:49.820 [256/740] Generating lib/distributor.sym_chk with a custom command (wrapped by meson to capture output) 00:02:49.820 [257/740] Linking target lib/librte_pci.so.23.0 00:02:49.820 [258/740] Linking target lib/librte_mempool.so.23.0 00:02:49.820 [259/740] Generating symbol file lib/librte_meter.so.23.0.p/librte_meter.so.23.0.symbols 00:02:49.820 [260/740] Linking target lib/librte_timer.so.23.0 00:02:49.820 [261/740] Generating symbol file lib/librte_rcu.so.23.0.p/librte_rcu.so.23.0.symbols 00:02:49.820 [262/740] Generating symbol file lib/librte_pci.so.23.0.p/librte_pci.so.23.0.symbols 00:02:49.820 [263/740] Linking target lib/librte_cfgfile.so.23.0 00:02:49.820 [264/740] Generating symbol file lib/librte_mempool.so.23.0.p/librte_mempool.so.23.0.symbols 00:02:49.820 [265/740] Linking target lib/librte_acl.so.23.0 00:02:50.078 [266/740] Linking target lib/librte_mbuf.so.23.0 00:02:50.078 [267/740] Generating symbol file lib/librte_timer.so.23.0.p/librte_timer.so.23.0.symbols 00:02:50.078 [268/740] Generating symbol file lib/librte_acl.so.23.0.p/librte_acl.so.23.0.symbols 00:02:50.078 [269/740] Generating symbol file lib/librte_mbuf.so.23.0.p/librte_mbuf.so.23.0.symbols 00:02:50.078 [270/740] Linking target lib/librte_net.so.23.0 00:02:50.337 [271/740] Generating symbol file lib/librte_net.so.23.0.p/librte_net.so.23.0.symbols 00:02:50.337 [272/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_ring.c.o 00:02:50.337 [273/740] Linking target lib/librte_cmdline.so.23.0 00:02:50.337 [274/740] Linking target lib/librte_hash.so.23.0 00:02:50.337 [275/740] Linking target lib/librte_bbdev.so.23.0 00:02:50.337 [276/740] Linking target lib/librte_compressdev.so.23.0 00:02:50.337 [277/740] Compiling C object lib/librte_efd.a.p/efd_rte_efd.c.o 00:02:50.337 [278/740] Linking static target lib/librte_efd.a 00:02:50.337 [279/740] Linking target lib/librte_distributor.so.23.0 00:02:50.337 [280/740] Generating lib/rte_eventdev_def with a custom command 00:02:50.337 [281/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_crypto_adapter.c.o 00:02:50.337 [282/740] Generating symbol file lib/librte_hash.so.23.0.p/librte_hash.so.23.0.symbols 00:02:50.337 [283/740] Generating lib/rte_eventdev_mingw with a custom command 00:02:50.337 [284/740] Generating lib/rte_gpudev_def with a custom command 00:02:50.337 [285/740] Generating lib/rte_gpudev_mingw with a custom command 00:02:50.596 [286/740] Generating lib/efd.sym_chk with a custom command (wrapped by meson to capture output) 00:02:50.596 [287/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:02:50.596 [288/740] Linking static target lib/librte_cryptodev.a 00:02:50.596 [289/740] Linking target lib/librte_efd.so.23.0 00:02:50.596 [290/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_tx_adapter.c.o 00:02:50.596 [291/740] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:50.854 [292/740] Linking target lib/librte_ethdev.so.23.0 00:02:50.854 [293/740] Generating symbol file lib/librte_ethdev.so.23.0.p/librte_ethdev.so.23.0.symbols 00:02:50.854 [294/740] Linking target lib/librte_metrics.so.23.0 00:02:50.854 [295/740] Compiling C object lib/librte_gro.a.p/gro_gro_tcp4.c.o 00:02:50.854 [296/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_timer_adapter.c.o 00:02:50.854 [297/740] Linking target lib/librte_bpf.so.23.0 00:02:50.854 [298/740] Compiling C object lib/librte_gpudev.a.p/gpudev_gpudev.c.o 00:02:50.854 [299/740] Generating symbol file lib/librte_metrics.so.23.0.p/librte_metrics.so.23.0.symbols 00:02:50.854 [300/740] Linking static target lib/librte_gpudev.a 00:02:50.854 [301/740] Compiling C object lib/librte_gro.a.p/gro_rte_gro.c.o 00:02:51.113 [302/740] Linking target lib/librte_bitratestats.so.23.0 00:02:51.113 [303/740] Generating lib/rte_gro_def with a custom command 00:02:51.113 [304/740] Generating lib/rte_gro_mingw with a custom command 00:02:51.113 [305/740] Compiling C object lib/librte_gro.a.p/gro_gro_udp4.c.o 00:02:51.113 [306/740] Generating symbol file lib/librte_bpf.so.23.0.p/librte_bpf.so.23.0.symbols 00:02:51.113 [307/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_eventdev.c.o 00:02:51.113 [308/740] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_tcp4.c.o 00:02:51.372 [309/740] Compiling C object lib/librte_gso.a.p/gso_gso_tcp4.c.o 00:02:51.372 [310/740] Compiling C object lib/librte_gso.a.p/gso_gso_udp4.c.o 00:02:51.372 [311/740] Generating lib/rte_gso_def with a custom command 00:02:51.372 [312/740] Generating lib/rte_gso_mingw with a custom command 00:02:51.372 [313/740] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_udp4.c.o 00:02:51.372 [314/740] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_udp4.c.o 00:02:51.372 [315/740] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_tcp4.c.o 00:02:51.372 [316/740] Linking static target lib/librte_gro.a 00:02:51.632 [317/740] Compiling C object lib/librte_gso.a.p/gso_rte_gso.c.o 00:02:51.632 [318/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_rx_adapter.c.o 00:02:51.632 [319/740] Linking static target lib/librte_eventdev.a 00:02:51.632 [320/740] Generating lib/gpudev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:51.632 [321/740] Linking target lib/librte_gpudev.so.23.0 00:02:51.632 [322/740] Compiling C object lib/librte_gso.a.p/gso_gso_common.c.o 00:02:51.632 [323/740] Generating lib/gro.sym_chk with a custom command (wrapped by meson to capture output) 00:02:51.632 [324/740] Linking static target lib/librte_gso.a 00:02:51.632 [325/740] Linking target lib/librte_gro.so.23.0 00:02:51.632 [326/740] Generating lib/rte_ip_frag_def with a custom command 00:02:51.632 [327/740] Generating lib/rte_ip_frag_mingw with a custom command 00:02:51.892 [328/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_reassembly.c.o 00:02:51.892 [329/740] Generating lib/gso.sym_chk with a custom command (wrapped by meson to capture output) 00:02:51.892 [330/740] Generating lib/rte_jobstats_def with a custom command 00:02:51.892 [331/740] Linking target lib/librte_gso.so.23.0 00:02:51.892 [332/740] Compiling C object lib/librte_jobstats.a.p/jobstats_rte_jobstats.c.o 00:02:51.892 [333/740] Linking static target lib/librte_jobstats.a 00:02:51.892 [334/740] Generating lib/rte_jobstats_mingw with a custom command 00:02:51.892 [335/740] Generating lib/rte_latencystats_def with a custom command 00:02:51.892 [336/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_reassembly.c.o 00:02:51.892 [337/740] Generating lib/rte_latencystats_mingw with a custom command 00:02:51.892 [338/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_fragmentation.c.o 00:02:51.892 [339/740] Generating lib/rte_lpm_def with a custom command 00:02:51.892 [340/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ip_frag_common.c.o 00:02:52.150 [341/740] Generating lib/rte_lpm_mingw with a custom command 00:02:52.151 [342/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_fragmentation.c.o 00:02:52.151 [343/740] Generating lib/jobstats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.151 [344/740] Linking target lib/librte_jobstats.so.23.0 00:02:52.151 [345/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_ip_frag_internal.c.o 00:02:52.151 [346/740] Linking static target lib/librte_ip_frag.a 00:02:52.151 [347/740] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.151 [348/740] Linking target lib/librte_cryptodev.so.23.0 00:02:52.409 [349/740] Generating symbol file lib/librte_cryptodev.so.23.0.p/librte_cryptodev.so.23.0.symbols 00:02:52.409 [350/740] Compiling C object lib/librte_latencystats.a.p/latencystats_rte_latencystats.c.o 00:02:52.409 [351/740] Linking static target lib/librte_latencystats.a 00:02:52.409 [352/740] Generating lib/ip_frag.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.409 [353/740] Compiling C object lib/librte_member.a.p/member_rte_member.c.o 00:02:52.409 [354/740] Linking target lib/librte_ip_frag.so.23.0 00:02:52.409 [355/740] Generating lib/rte_member_def with a custom command 00:02:52.409 [356/740] Generating lib/rte_member_mingw with a custom command 00:02:52.409 [357/740] Compiling C object lib/member/libsketch_avx512_tmp.a.p/rte_member_sketch_avx512.c.o 00:02:52.409 [358/740] Linking static target lib/member/libsketch_avx512_tmp.a 00:02:52.409 [359/740] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm.c.o 00:02:52.409 [360/740] Generating lib/rte_pcapng_def with a custom command 00:02:52.409 [361/740] Generating lib/rte_pcapng_mingw with a custom command 00:02:52.668 [362/740] Generating symbol file lib/librte_ip_frag.so.23.0.p/librte_ip_frag.so.23.0.symbols 00:02:52.668 [363/740] Generating lib/latencystats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.668 [364/740] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:02:52.668 [365/740] Linking target lib/librte_latencystats.so.23.0 00:02:52.668 [366/740] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:02:52.668 [367/740] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:02:52.668 [368/740] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:02:52.927 [369/740] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm6.c.o 00:02:52.927 [370/740] Linking static target lib/librte_lpm.a 00:02:52.927 [371/740] Compiling C object lib/librte_member.a.p/member_rte_member_ht.c.o 00:02:52.927 [372/740] Compiling C object lib/librte_member.a.p/member_rte_member_vbf.c.o 00:02:52.927 [373/740] Compiling C object lib/librte_power.a.p/power_rte_power_empty_poll.c.o 00:02:52.927 [374/740] Generating lib/rte_power_def with a custom command 00:02:52.927 [375/740] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:02:52.927 [376/740] Generating lib/rte_power_mingw with a custom command 00:02:52.927 [377/740] Generating lib/rte_rawdev_def with a custom command 00:02:52.927 [378/740] Generating lib/rte_rawdev_mingw with a custom command 00:02:53.185 [379/740] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:02:53.185 [380/740] Generating lib/rte_regexdev_def with a custom command 00:02:53.185 [381/740] Generating lib/eventdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:53.185 [382/740] Generating lib/rte_regexdev_mingw with a custom command 00:02:53.185 [383/740] Compiling C object lib/librte_pcapng.a.p/pcapng_rte_pcapng.c.o 00:02:53.185 [384/740] Linking static target lib/librte_pcapng.a 00:02:53.185 [385/740] Linking target lib/librte_eventdev.so.23.0 00:02:53.185 [386/740] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:02:53.185 [387/740] Generating lib/lpm.sym_chk with a custom command (wrapped by meson to capture output) 00:02:53.185 [388/740] Generating lib/rte_dmadev_def with a custom command 00:02:53.185 [389/740] Generating lib/rte_dmadev_mingw with a custom command 00:02:53.185 [390/740] Linking target lib/librte_lpm.so.23.0 00:02:53.185 [391/740] Generating symbol file lib/librte_eventdev.so.23.0.p/librte_eventdev.so.23.0.symbols 00:02:53.185 [392/740] Compiling C object lib/librte_rawdev.a.p/rawdev_rte_rawdev.c.o 00:02:53.185 [393/740] Linking static target lib/librte_rawdev.a 00:02:53.185 [394/740] Compiling C object lib/librte_power.a.p/power_rte_power_intel_uncore.c.o 00:02:53.444 [395/740] Generating lib/rte_rib_def with a custom command 00:02:53.444 [396/740] Generating lib/rte_rib_mingw with a custom command 00:02:53.444 [397/740] Generating symbol file lib/librte_lpm.so.23.0.p/librte_lpm.so.23.0.symbols 00:02:53.444 [398/740] Generating lib/rte_reorder_def with a custom command 00:02:53.444 [399/740] Generating lib/pcapng.sym_chk with a custom command (wrapped by meson to capture output) 00:02:53.444 [400/740] Generating lib/rte_reorder_mingw with a custom command 00:02:53.444 [401/740] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:02:53.444 [402/740] Linking static target lib/librte_power.a 00:02:53.444 [403/740] Linking target lib/librte_pcapng.so.23.0 00:02:53.444 [404/740] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:02:53.444 [405/740] Linking static target lib/librte_dmadev.a 00:02:53.444 [406/740] Compiling C object lib/librte_regexdev.a.p/regexdev_rte_regexdev.c.o 00:02:53.444 [407/740] Linking static target lib/librte_regexdev.a 00:02:53.444 [408/740] Generating symbol file lib/librte_pcapng.so.23.0.p/librte_pcapng.so.23.0.symbols 00:02:53.703 [409/740] Compiling C object lib/librte_sched.a.p/sched_rte_red.c.o 00:02:53.703 [410/740] Generating lib/rawdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:53.703 [411/740] Linking target lib/librte_rawdev.so.23.0 00:02:53.703 [412/740] Compiling C object lib/librte_rib.a.p/rib_rte_rib.c.o 00:02:53.703 [413/740] Compiling C object lib/librte_sched.a.p/sched_rte_approx.c.o 00:02:53.703 [414/740] Generating lib/rte_sched_def with a custom command 00:02:53.703 [415/740] Compiling C object lib/librte_member.a.p/member_rte_member_sketch.c.o 00:02:53.703 [416/740] Generating lib/rte_sched_mingw with a custom command 00:02:53.703 [417/740] Compiling C object lib/librte_sched.a.p/sched_rte_pie.c.o 00:02:53.703 [418/740] Linking static target lib/librte_member.a 00:02:53.703 [419/740] Generating lib/rte_security_def with a custom command 00:02:53.703 [420/740] Generating lib/rte_security_mingw with a custom command 00:02:53.962 [421/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack_std.c.o 00:02:53.962 [422/740] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:02:53.962 [423/740] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:53.962 [424/740] Linking static target lib/librte_reorder.a 00:02:53.962 [425/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack.c.o 00:02:53.962 [426/740] Linking target lib/librte_dmadev.so.23.0 00:02:53.962 [427/740] Generating lib/rte_stack_def with a custom command 00:02:53.962 [428/740] Generating lib/rte_stack_mingw with a custom command 00:02:53.962 [429/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack_lf.c.o 00:02:53.962 [430/740] Linking static target lib/librte_stack.a 00:02:53.962 [431/740] Compiling C object lib/librte_rib.a.p/rib_rte_rib6.c.o 00:02:53.962 [432/740] Linking static target lib/librte_rib.a 00:02:53.962 [433/740] Generating symbol file lib/librte_dmadev.so.23.0.p/librte_dmadev.so.23.0.symbols 00:02:54.237 [434/740] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:02:54.237 [435/740] Generating lib/member.sym_chk with a custom command (wrapped by meson to capture output) 00:02:54.237 [436/740] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:02:54.237 [437/740] Generating lib/regexdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:54.237 [438/740] Generating lib/stack.sym_chk with a custom command (wrapped by meson to capture output) 00:02:54.237 [439/740] Linking target lib/librte_member.so.23.0 00:02:54.237 [440/740] Linking target lib/librte_reorder.so.23.0 00:02:54.237 [441/740] Linking target lib/librte_stack.so.23.0 00:02:54.237 [442/740] Linking target lib/librte_regexdev.so.23.0 00:02:54.237 [443/740] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:02:54.237 [444/740] Linking target lib/librte_power.so.23.0 00:02:54.237 [445/740] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:02:54.237 [446/740] Linking static target lib/librte_security.a 00:02:54.237 [447/740] Generating lib/rib.sym_chk with a custom command (wrapped by meson to capture output) 00:02:54.496 [448/740] Linking target lib/librte_rib.so.23.0 00:02:54.496 [449/740] Generating symbol file lib/librte_rib.so.23.0.p/librte_rib.so.23.0.symbols 00:02:54.496 [450/740] Generating lib/rte_vhost_def with a custom command 00:02:54.496 [451/740] Generating lib/rte_vhost_mingw with a custom command 00:02:54.496 [452/740] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:02:54.496 [453/740] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:02:54.756 [454/740] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:02:54.756 [455/740] Linking target lib/librte_security.so.23.0 00:02:54.756 [456/740] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:02:54.756 [457/740] Compiling C object lib/librte_sched.a.p/sched_rte_sched.c.o 00:02:54.756 [458/740] Linking static target lib/librte_sched.a 00:02:54.756 [459/740] Generating symbol file lib/librte_security.so.23.0.p/librte_security.so.23.0.symbols 00:02:55.016 [460/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ses.c.o 00:02:55.016 [461/740] Generating lib/sched.sym_chk with a custom command (wrapped by meson to capture output) 00:02:55.016 [462/740] Linking target lib/librte_sched.so.23.0 00:02:55.276 [463/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:02:55.276 [464/740] Compiling C object lib/librte_ipsec.a.p/ipsec_sa.c.o 00:02:55.276 [465/740] Generating lib/rte_ipsec_def with a custom command 00:02:55.276 [466/740] Generating lib/rte_ipsec_mingw with a custom command 00:02:55.276 [467/740] Generating symbol file lib/librte_sched.so.23.0.p/librte_sched.so.23.0.symbols 00:02:55.276 [468/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:02:55.276 [469/740] Compiling C object lib/librte_fib.a.p/fib_rte_fib.c.o 00:02:55.536 [470/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_sad.c.o 00:02:55.536 [471/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_telemetry.c.o 00:02:55.536 [472/740] Generating lib/rte_fib_def with a custom command 00:02:55.536 [473/740] Generating lib/rte_fib_mingw with a custom command 00:02:55.536 [474/740] Compiling C object lib/librte_fib.a.p/fib_rte_fib6.c.o 00:02:55.800 [475/740] Compiling C object lib/librte_fib.a.p/fib_dir24_8_avx512.c.o 00:02:55.800 [476/740] Compiling C object lib/librte_fib.a.p/fib_trie_avx512.c.o 00:02:56.097 [477/740] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_outb.c.o 00:02:56.097 [478/740] Compiling C object lib/librte_fib.a.p/fib_trie.c.o 00:02:56.097 [479/740] Compiling C object lib/librte_fib.a.p/fib_dir24_8.c.o 00:02:56.097 [480/740] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_inb.c.o 00:02:56.097 [481/740] Linking static target lib/librte_fib.a 00:02:56.097 [482/740] Linking static target lib/librte_ipsec.a 00:02:56.097 [483/740] Compiling C object lib/librte_port.a.p/port_rte_port_ethdev.c.o 00:02:56.097 [484/740] Compiling C object lib/librte_port.a.p/port_rte_port_fd.c.o 00:02:56.372 [485/740] Compiling C object lib/librte_port.a.p/port_rte_port_frag.c.o 00:02:56.372 [486/740] Compiling C object lib/librte_port.a.p/port_rte_port_sched.c.o 00:02:56.372 [487/740] Generating lib/fib.sym_chk with a custom command (wrapped by meson to capture output) 00:02:56.372 [488/740] Linking target lib/librte_fib.so.23.0 00:02:56.372 [489/740] Generating lib/ipsec.sym_chk with a custom command (wrapped by meson to capture output) 00:02:56.372 [490/740] Compiling C object lib/librte_port.a.p/port_rte_port_ras.c.o 00:02:56.372 [491/740] Linking target lib/librte_ipsec.so.23.0 00:02:56.941 [492/740] Compiling C object lib/librte_port.a.p/port_rte_port_sym_crypto.c.o 00:02:56.941 [493/740] Generating lib/rte_port_def with a custom command 00:02:56.941 [494/740] Generating lib/rte_port_mingw with a custom command 00:02:56.941 [495/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ethdev.c.o 00:02:56.941 [496/740] Compiling C object lib/librte_port.a.p/port_rte_port_source_sink.c.o 00:02:56.941 [497/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_fd.c.o 00:02:56.941 [498/740] Generating lib/rte_pdump_def with a custom command 00:02:56.941 [499/740] Generating lib/rte_pdump_mingw with a custom command 00:02:57.200 [500/740] Compiling C object lib/librte_port.a.p/port_rte_port_eventdev.c.o 00:02:57.200 [501/740] Compiling C object lib/librte_table.a.p/table_rte_swx_keycmp.c.o 00:02:57.200 [502/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_source_sink.c.o 00:02:57.200 [503/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_learner.c.o 00:02:57.200 [504/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_em.c.o 00:02:57.200 [505/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ring.c.o 00:02:57.460 [506/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_selector.c.o 00:02:57.460 [507/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_wm.c.o 00:02:57.460 [508/740] Compiling C object lib/librte_port.a.p/port_rte_port_ring.c.o 00:02:57.460 [509/740] Linking static target lib/librte_port.a 00:02:57.718 [510/740] Compiling C object lib/librte_table.a.p/table_rte_table_array.c.o 00:02:57.718 [511/740] Compiling C object lib/librte_table.a.p/table_rte_table_acl.c.o 00:02:57.718 [512/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_cuckoo.c.o 00:02:57.718 [513/740] Compiling C object lib/librte_pdump.a.p/pdump_rte_pdump.c.o 00:02:57.718 [514/740] Linking static target lib/librte_pdump.a 00:02:57.976 [515/740] Generating lib/port.sym_chk with a custom command (wrapped by meson to capture output) 00:02:57.976 [516/740] Generating lib/pdump.sym_chk with a custom command (wrapped by meson to capture output) 00:02:57.976 [517/740] Linking target lib/librte_port.so.23.0 00:02:57.976 [518/740] Compiling C object lib/librte_table.a.p/table_rte_table_lpm.c.o 00:02:57.976 [519/740] Linking target lib/librte_pdump.so.23.0 00:02:58.235 [520/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_ext.c.o 00:02:58.235 [521/740] Generating lib/rte_table_def with a custom command 00:02:58.235 [522/740] Generating symbol file lib/librte_port.so.23.0.p/librte_port.so.23.0.symbols 00:02:58.235 [523/740] Generating lib/rte_table_mingw with a custom command 00:02:58.235 [524/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key8.c.o 00:02:58.235 [525/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key16.c.o 00:02:58.494 [526/740] Compiling C object lib/librte_table.a.p/table_rte_table_stub.c.o 00:02:58.494 [527/740] Compiling C object lib/librte_table.a.p/table_rte_table_lpm_ipv6.c.o 00:02:58.494 [528/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_lru.c.o 00:02:58.494 [529/740] Generating lib/rte_pipeline_def with a custom command 00:02:58.494 [530/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_port_in_action.c.o 00:02:58.494 [531/740] Generating lib/rte_pipeline_mingw with a custom command 00:02:58.753 [532/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key32.c.o 00:02:58.753 [533/740] Linking static target lib/librte_table.a 00:02:58.753 [534/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_pipeline.c.o 00:02:59.011 [535/740] Compiling C object lib/librte_graph.a.p/graph_node.c.o 00:02:59.012 [536/740] Compiling C object lib/librte_graph.a.p/graph_graph.c.o 00:02:59.270 [537/740] Generating lib/table.sym_chk with a custom command (wrapped by meson to capture output) 00:02:59.270 [538/740] Linking target lib/librte_table.so.23.0 00:02:59.270 [539/740] Compiling C object lib/librte_graph.a.p/graph_graph_ops.c.o 00:02:59.270 [540/740] Generating symbol file lib/librte_table.so.23.0.p/librte_table.so.23.0.symbols 00:02:59.270 [541/740] Generating lib/rte_graph_def with a custom command 00:02:59.270 [542/740] Compiling C object lib/librte_graph.a.p/graph_graph_debug.c.o 00:02:59.270 [543/740] Generating lib/rte_graph_mingw with a custom command 00:02:59.529 [544/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_ctl.c.o 00:02:59.529 [545/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:02:59.529 [546/740] Compiling C object lib/librte_graph.a.p/graph_graph_stats.c.o 00:02:59.529 [547/740] Compiling C object lib/librte_graph.a.p/graph_graph_populate.c.o 00:02:59.529 [548/740] Linking static target lib/librte_graph.a 00:02:59.791 [549/740] Compiling C object lib/librte_node.a.p/node_ethdev_ctrl.c.o 00:02:59.791 [550/740] Compiling C object lib/librte_node.a.p/node_ethdev_rx.c.o 00:02:59.791 [551/740] Compiling C object lib/librte_node.a.p/node_null.c.o 00:02:59.791 [552/740] Compiling C object lib/librte_node.a.p/node_ethdev_tx.c.o 00:03:00.050 [553/740] Compiling C object lib/librte_node.a.p/node_log.c.o 00:03:00.050 [554/740] Generating lib/rte_node_def with a custom command 00:03:00.050 [555/740] Generating lib/rte_node_mingw with a custom command 00:03:00.310 [556/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline_spec.c.o 00:03:00.310 [557/740] Generating lib/graph.sym_chk with a custom command (wrapped by meson to capture output) 00:03:00.310 [558/740] Compiling C object lib/librte_node.a.p/node_pkt_drop.c.o 00:03:00.310 [559/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:03:00.310 [560/740] Linking target lib/librte_graph.so.23.0 00:03:00.310 [561/740] Compiling C object lib/librte_node.a.p/node_ip4_lookup.c.o 00:03:00.310 [562/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:03:00.310 [563/740] Generating symbol file lib/librte_graph.so.23.0.p/librte_graph.so.23.0.symbols 00:03:00.310 [564/740] Generating drivers/rte_bus_pci_def with a custom command 00:03:00.310 [565/740] Generating drivers/rte_bus_pci_mingw with a custom command 00:03:00.569 [566/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:03:00.569 [567/740] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:03:00.569 [568/740] Compiling C object lib/librte_node.a.p/node_ip4_rewrite.c.o 00:03:00.569 [569/740] Compiling C object lib/librte_node.a.p/node_pkt_cls.c.o 00:03:00.569 [570/740] Generating drivers/rte_bus_vdev_def with a custom command 00:03:00.569 [571/740] Linking static target lib/librte_node.a 00:03:00.569 [572/740] Generating drivers/rte_bus_vdev_mingw with a custom command 00:03:00.569 [573/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:03:00.569 [574/740] Generating drivers/rte_mempool_ring_def with a custom command 00:03:00.569 [575/740] Generating drivers/rte_mempool_ring_mingw with a custom command 00:03:00.569 [576/740] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:03:00.569 [577/740] Linking static target drivers/libtmp_rte_bus_vdev.a 00:03:00.569 [578/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:03:00.828 [579/740] Generating lib/node.sym_chk with a custom command (wrapped by meson to capture output) 00:03:00.828 [580/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:03:00.828 [581/740] Linking static target drivers/libtmp_rte_bus_pci.a 00:03:00.828 [582/740] Linking target lib/librte_node.so.23.0 00:03:00.828 [583/740] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:03:00.828 [584/740] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:03:00.828 [585/740] Linking static target drivers/librte_bus_vdev.a 00:03:00.828 [586/740] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:03:00.828 [587/740] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:03:00.828 [588/740] Linking static target drivers/librte_bus_pci.a 00:03:01.087 [589/740] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:03:01.087 [590/740] Compiling C object drivers/librte_bus_pci.so.23.0.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:03:01.087 [591/740] Compiling C object drivers/librte_bus_vdev.so.23.0.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:03:01.087 [592/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_diag.c.o 00:03:01.087 [593/740] Linking target drivers/librte_bus_vdev.so.23.0 00:03:01.347 [594/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_dcb.c.o 00:03:01.347 [595/740] Generating symbol file drivers/librte_bus_vdev.so.23.0.p/librte_bus_vdev.so.23.0.symbols 00:03:01.347 [596/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_adminq.c.o 00:03:01.347 [597/740] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:03:01.347 [598/740] Linking target drivers/librte_bus_pci.so.23.0 00:03:01.347 [599/740] Generating symbol file drivers/librte_bus_pci.so.23.0.p/librte_bus_pci.so.23.0.symbols 00:03:01.347 [600/740] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:03:01.347 [601/740] Linking static target drivers/libtmp_rte_mempool_ring.a 00:03:01.605 [602/740] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:03:01.605 [603/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_hmc.c.o 00:03:01.605 [604/740] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:03:01.605 [605/740] Linking static target drivers/librte_mempool_ring.a 00:03:01.605 [606/740] Compiling C object drivers/librte_mempool_ring.so.23.0.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:03:01.605 [607/740] Linking target drivers/librte_mempool_ring.so.23.0 00:03:01.864 [608/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_lan_hmc.c.o 00:03:02.122 [609/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_nvm.c.o 00:03:02.381 [610/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_common.c.o 00:03:02.381 [611/740] Linking static target drivers/net/i40e/base/libi40e_base.a 00:03:02.381 [612/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_pf.c.o 00:03:02.640 [613/740] Compiling C object drivers/net/i40e/libi40e_avx512_lib.a.p/i40e_rxtx_vec_avx512.c.o 00:03:02.640 [614/740] Linking static target drivers/net/i40e/libi40e_avx512_lib.a 00:03:02.899 [615/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_tm.c.o 00:03:02.899 [616/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_fdir.c.o 00:03:03.158 [617/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_hash.c.o 00:03:03.158 [618/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_flow.c.o 00:03:03.158 [619/740] Generating drivers/rte_net_i40e_def with a custom command 00:03:03.158 [620/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_vf_representor.c.o 00:03:03.158 [621/740] Generating drivers/rte_net_i40e_mingw with a custom command 00:03:03.728 [622/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline.c.o 00:03:03.988 [623/740] Compiling C object app/dpdk-dumpcap.p/dumpcap_main.c.o 00:03:04.249 [624/740] Compiling C object app/dpdk-pdump.p/pdump_main.c.o 00:03:04.249 [625/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_sse.c.o 00:03:04.249 [626/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_rte_pmd_i40e.c.o 00:03:04.249 [627/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_main.c.o 00:03:04.249 [628/740] Compiling C object app/dpdk-proc-info.p/proc-info_main.c.o 00:03:04.509 [629/740] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_commands.c.o 00:03:04.509 [630/740] Compiling C object app/dpdk-test-acl.p/test-acl_main.c.o 00:03:04.509 [631/740] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_cmdline_test.c.o 00:03:04.509 [632/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx.c.o 00:03:04.769 [633/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_avx2.c.o 00:03:05.029 [634/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_options_parse.c.o 00:03:05.029 [635/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev.c.o 00:03:05.029 [636/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_vector.c.o 00:03:05.290 [637/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_ethdev.c.o 00:03:05.290 [638/740] Linking static target drivers/libtmp_rte_net_i40e.a 00:03:05.290 [639/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_common.c.o 00:03:05.290 [640/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_cyclecount.c.o 00:03:05.549 [641/740] Generating drivers/rte_net_i40e.pmd.c with a custom command 00:03:05.550 [642/740] Compiling C object drivers/librte_net_i40e.a.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:03:05.550 [643/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_throughput.c.o 00:03:05.550 [644/740] Linking static target drivers/librte_net_i40e.a 00:03:05.550 [645/740] Compiling C object drivers/librte_net_i40e.so.23.0.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:03:05.550 [646/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_main.c.o 00:03:05.550 [647/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_verify.c.o 00:03:05.808 [648/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_ops.c.o 00:03:05.808 [649/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_common.c.o 00:03:06.066 [650/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_options_parsing.c.o 00:03:06.066 [651/740] Generating drivers/rte_net_i40e.sym_chk with a custom command (wrapped by meson to capture output) 00:03:06.066 [652/740] Linking target drivers/librte_net_i40e.so.23.0 00:03:06.066 [653/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vectors.c.o 00:03:06.325 [654/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_pmd_cyclecount.c.o 00:03:06.325 [655/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vector_parsing.c.o 00:03:06.325 [656/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_latency.c.o 00:03:06.325 [657/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_test.c.o 00:03:06.325 [658/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_throughput.c.o 00:03:06.325 [659/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_parser.c.o 00:03:06.584 [660/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_main.c.o 00:03:06.584 [661/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_main.c.o 00:03:06.584 [662/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_options.c.o 00:03:06.584 [663/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_verify.c.o 00:03:06.850 [664/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_common.c.o 00:03:06.850 [665/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_atq.c.o 00:03:07.117 [666/740] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:03:07.117 [667/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_queue.c.o 00:03:07.117 [668/740] Linking static target lib/librte_vhost.a 00:03:07.377 [669/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_atq.c.o 00:03:07.639 [670/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_queue.c.o 00:03:07.639 [671/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_atq.c.o 00:03:07.639 [672/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_common.c.o 00:03:07.898 [673/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_queue.c.o 00:03:07.898 [674/740] Compiling C object app/dpdk-test-fib.p/test-fib_main.c.o 00:03:07.898 [675/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_flow_gen.c.o 00:03:07.898 [676/740] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:03:07.898 [677/740] Linking target lib/librte_vhost.so.23.0 00:03:08.157 [678/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_actions_gen.c.o 00:03:08.157 [679/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_items_gen.c.o 00:03:08.157 [680/740] Compiling C object app/dpdk-test-gpudev.p/test-gpudev_main.c.o 00:03:08.157 [681/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_config.c.o 00:03:08.415 [682/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_common.c.o 00:03:08.415 [683/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_init.c.o 00:03:08.415 [684/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_main.c.o 00:03:08.415 [685/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_acl.c.o 00:03:08.674 [686/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm.c.o 00:03:08.674 [687/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_hash.c.o 00:03:08.674 [688/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm_ipv6.c.o 00:03:08.674 [689/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_main.c.o 00:03:08.674 [690/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_stub.c.o 00:03:08.933 [691/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_perf.c.o 00:03:08.933 [692/740] Compiling C object app/dpdk-testpmd.p/test-pmd_5tswap.c.o 00:03:08.933 [693/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_runtime.c.o 00:03:08.933 [694/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmd_flex_item.c.o 00:03:09.192 [695/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_mtr.c.o 00:03:09.192 [696/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_tm.c.o 00:03:09.450 [697/740] Compiling C object app/dpdk-testpmd.p/test-pmd_ieee1588fwd.c.o 00:03:09.450 [698/740] Compiling C object app/dpdk-testpmd.p/test-pmd_flowgen.c.o 00:03:09.725 [699/740] Compiling C object app/dpdk-testpmd.p/test-pmd_icmpecho.c.o 00:03:09.725 [700/740] Compiling C object app/dpdk-testpmd.p/test-pmd_iofwd.c.o 00:03:09.725 [701/740] Compiling C object app/dpdk-testpmd.p/test-pmd_macfwd.c.o 00:03:10.294 [702/740] Compiling C object app/dpdk-testpmd.p/test-pmd_macswap.c.o 00:03:10.294 [703/740] Compiling C object app/dpdk-testpmd.p/test-pmd_csumonly.c.o 00:03:10.294 [704/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline.c.o 00:03:10.294 [705/740] Compiling C object app/dpdk-testpmd.p/test-pmd_rxonly.c.o 00:03:10.294 [706/740] Compiling C object app/dpdk-testpmd.p/test-pmd_shared_rxq_fwd.c.o 00:03:10.294 [707/740] Compiling C object app/dpdk-testpmd.p/test-pmd_parameters.c.o 00:03:10.554 [708/740] Compiling C object app/dpdk-testpmd.p/test-pmd_bpf_cmd.c.o 00:03:10.814 [709/740] Compiling C object app/dpdk-testpmd.p/test-pmd_util.c.o 00:03:11.074 [710/740] Compiling C object app/dpdk-testpmd.p/test-pmd_config.c.o 00:03:11.074 [711/740] Compiling C object app/dpdk-testpmd.p/.._drivers_net_i40e_i40e_testpmd.c.o 00:03:11.074 [712/740] Compiling C object app/dpdk-test-regex.p/test-regex_main.c.o 00:03:11.074 [713/740] Compiling C object app/dpdk-test-sad.p/test-sad_main.c.o 00:03:11.335 [714/740] Compiling C object app/dpdk-testpmd.p/test-pmd_txonly.c.o 00:03:11.335 [715/740] Compiling C object app/dpdk-test-security-perf.p/test-security-perf_test_security_perf.c.o 00:03:11.335 [716/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_flow.c.o 00:03:11.594 [717/740] Compiling C object app/dpdk-testpmd.p/test-pmd_noisy_vnf.c.o 00:03:11.594 [718/740] Compiling C object app/dpdk-testpmd.p/test-pmd_testpmd.c.o 00:03:11.852 [719/740] Compiling C object app/dpdk-test-security-perf.p/test_test_cryptodev_security_ipsec.c.o 00:03:12.789 [720/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_table_action.c.o 00:03:12.789 [721/740] Linking static target lib/librte_pipeline.a 00:03:13.047 [722/740] Linking target app/dpdk-dumpcap 00:03:13.047 [723/740] Linking target app/dpdk-pdump 00:03:13.047 [724/740] Linking target app/dpdk-test-acl 00:03:13.047 [725/740] Linking target app/dpdk-test-compress-perf 00:03:13.047 [726/740] Linking target app/dpdk-test-bbdev 00:03:13.047 [727/740] Linking target app/dpdk-test-crypto-perf 00:03:13.047 [728/740] Linking target app/dpdk-test-eventdev 00:03:13.047 [729/740] Linking target app/dpdk-test-cmdline 00:03:13.047 [730/740] Linking target app/dpdk-proc-info 00:03:13.307 [731/740] Linking target app/dpdk-test-pipeline 00:03:13.307 [732/740] Linking target app/dpdk-test-regex 00:03:13.307 [733/740] Linking target app/dpdk-test-sad 00:03:13.307 [734/740] Linking target app/dpdk-test-fib 00:03:13.307 [735/740] Linking target app/dpdk-test-gpudev 00:03:13.307 [736/740] Linking target app/dpdk-test-security-perf 00:03:13.307 [737/740] Linking target app/dpdk-testpmd 00:03:13.307 [738/740] Linking target app/dpdk-test-flow-perf 00:03:17.527 [739/740] Generating lib/pipeline.sym_chk with a custom command (wrapped by meson to capture output) 00:03:17.787 [740/740] Linking target lib/librte_pipeline.so.23.0 00:03:17.787 18:53:35 build_native_dpdk -- common/autobuild_common.sh@201 -- $ uname -s 00:03:17.787 18:53:35 build_native_dpdk -- common/autobuild_common.sh@201 -- $ [[ Linux == \F\r\e\e\B\S\D ]] 00:03:17.787 18:53:35 build_native_dpdk -- common/autobuild_common.sh@214 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 install 00:03:17.787 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:03:17.787 [0/1] Installing files. 00:03:18.048 Installing subdir /home/vagrant/spdk_repo/dpdk/examples to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/README to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/dummy.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t1.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t2.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t3.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/common/pkt_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/common/altivec/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/altivec 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/common/neon/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/neon 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/common/sse/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/sse 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/dmafwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_tx.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_aes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ccm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_cmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ecdsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_gcm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_hmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_rsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_sha.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_tdes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_xts.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.048 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/flow_classify.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/ipv4_rules_file.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/flow_blocks.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/kni.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/kni.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/firewall.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow_crypto.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/kni.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route_ecmp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/rss.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/tap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep0.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep1.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipip.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_process.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/rt.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp4.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp6.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/bypass_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs_secgw.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/data_rxtx.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/linux_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/load_env.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/run_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_ipv6opts.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_null_header_reconstruct.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/l2fwd-cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:18.049 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl_scalar.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_sequential.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_fib.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_route.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/client.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/shared 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/ntb_fwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ethdev.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_group_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_routing_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/packet.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/pcap.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/ptpclient.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/app_thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:18.050 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cmdline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_ov.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_pie.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_red.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/stats.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/node/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/node 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/node/node.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/node 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/shared 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/basicfwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/vdpa_blk_compact.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/virtio_net.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk_spec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk_compat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_nop.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_x86.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:03:18.051 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:03:18.051 Installing lib/librte_kvargs.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.051 Installing lib/librte_kvargs.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.051 Installing lib/librte_telemetry.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.051 Installing lib/librte_telemetry.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.051 Installing lib/librte_eal.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.051 Installing lib/librte_eal.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.051 Installing lib/librte_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.051 Installing lib/librte_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.051 Installing lib/librte_rcu.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.051 Installing lib/librte_rcu.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.051 Installing lib/librte_mempool.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.051 Installing lib/librte_mempool.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.051 Installing lib/librte_mbuf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.051 Installing lib/librte_mbuf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.051 Installing lib/librte_net.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_net.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_meter.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_meter.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_ethdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_ethdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_cmdline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_cmdline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_metrics.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_metrics.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_hash.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_hash.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_timer.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_timer.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_acl.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_acl.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_bbdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_bbdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_bitratestats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_bitratestats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_bpf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_bpf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_cfgfile.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_cfgfile.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_compressdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_compressdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_cryptodev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_cryptodev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_distributor.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_distributor.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_efd.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_efd.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_eventdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_eventdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_gpudev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_gpudev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_gro.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_gro.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_gso.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_gso.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_ip_frag.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_ip_frag.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_jobstats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_jobstats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_latencystats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_latencystats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_lpm.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_lpm.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_member.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_member.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_pcapng.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_pcapng.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_power.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_power.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_rawdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_rawdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_regexdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_regexdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_dmadev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_dmadev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_rib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_rib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_reorder.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_reorder.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_sched.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_sched.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_security.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_security.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_stack.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_stack.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_vhost.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_vhost.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_ipsec.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_ipsec.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_fib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_fib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_port.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_port.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_pdump.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_pdump.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_table.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_table.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_pipeline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_pipeline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_graph.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_graph.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_node.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing lib/librte_node.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing drivers/librte_bus_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing drivers/librte_bus_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:03:18.332 Installing drivers/librte_bus_vdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing drivers/librte_bus_vdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:03:18.332 Installing drivers/librte_mempool_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing drivers/librte_mempool_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:03:18.332 Installing drivers/librte_net_i40e.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.332 Installing drivers/librte_net_i40e.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:03:18.332 Installing app/dpdk-dumpcap to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-pdump to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-proc-info to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-test-acl to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-test-bbdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-test-cmdline to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-test-compress-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-test-crypto-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-test-eventdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-test-fib to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-test-flow-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-test-gpudev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-test-pipeline to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-testpmd to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-test-regex to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-test-sad to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing app/dpdk-test-security-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/config/rte_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/kvargs/rte_kvargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/telemetry/rte_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rtm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_alarm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitmap.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_branch_prediction.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bus.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_class.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_compat.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_debug.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_dev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_devargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_memconfig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_errno.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_epoll.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_fbarray.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hexdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hypervisor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_interrupts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_keepalive.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_launch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_log.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_malloc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_mcslock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memory.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memzone.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_feature_defs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_features.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_per_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pflock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_random.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_reciprocal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqcount.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service_component.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_string_fns.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_tailq.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_thread.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_ticketlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_time.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point_register.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.332 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_uuid.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_version.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_vfio.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/linux/include/rte_os.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_c11_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_generic_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_zc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/rcu/rte_rcu_qsbr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_ptype.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_pool_ops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_dyn.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ip.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_tcp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_udp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_esp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_sctp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_icmp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_arp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ether.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_macsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_vxlan.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gre.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gtp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_mpls.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_higig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ecpri.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_geneve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_l2tpv2.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ppp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/meter/rte_meter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_cman.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_dev_info.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_eth_ctrl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/pci/rte_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_num.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_ipaddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_etheraddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_string.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_rdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_vt100.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_socket.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_cirbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_portlist.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_fbk_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_jhash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_sw.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_x86_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/timer/rte_timer.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl_osdep.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_op.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/bitratestats/rte_bitrate.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/bpf_def.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cfgfile/rte_cfgfile.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_compressdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_comp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_sym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_asym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/distributor/rte_distributor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/efd/rte_efd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_crypto_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_rx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_tx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_timer_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/gpudev/rte_gpudev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/gro/rte_gro.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/gso/rte_gso.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ip_frag/rte_ip_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/jobstats/rte_jobstats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/latencystats/rte_latencystats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_scalar.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/member/rte_member.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/pcapng/rte_pcapng.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_empty_poll.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_intel_uncore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_pmd_mgmt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_guest_channel.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/reorder/rte_reorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_approx.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_red.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_pie.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_std.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_c11.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_stubs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vdpa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_async.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sad.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_group.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ras.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sym_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/pdump/rte_pdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_em.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.333 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_learner.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.334 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_selector.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.334 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_wm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.334 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.334 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.334 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_array.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.334 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.334 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_cuckoo.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.334 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.334 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.334 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm_ipv6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.334 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_stub.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.334 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.334 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_port_in_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_table_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_extern.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_ctl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_worker.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_ip4_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_eth_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/pci/rte_bus_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/vdev/rte_bus_vdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/drivers/net/i40e/rte_pmd_i40e.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-devbind.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-pmdinfo.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-telemetry.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-hugepages.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/rte_build_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk-libs.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:03:18.594 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:03:18.594 Installing symlink pointing to librte_kvargs.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so.23 00:03:18.594 Installing symlink pointing to librte_kvargs.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so 00:03:18.594 Installing symlink pointing to librte_telemetry.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so.23 00:03:18.594 Installing symlink pointing to librte_telemetry.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so 00:03:18.594 Installing symlink pointing to librte_eal.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so.23 00:03:18.594 Installing symlink pointing to librte_eal.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so 00:03:18.594 Installing symlink pointing to librte_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so.23 00:03:18.594 Installing symlink pointing to librte_ring.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so 00:03:18.594 Installing symlink pointing to librte_rcu.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so.23 00:03:18.594 Installing symlink pointing to librte_rcu.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so 00:03:18.594 Installing symlink pointing to librte_mempool.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so.23 00:03:18.594 Installing symlink pointing to librte_mempool.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so 00:03:18.594 Installing symlink pointing to librte_mbuf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so.23 00:03:18.594 Installing symlink pointing to librte_mbuf.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so 00:03:18.594 Installing symlink pointing to librte_net.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so.23 00:03:18.594 Installing symlink pointing to librte_net.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so 00:03:18.594 Installing symlink pointing to librte_meter.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so.23 00:03:18.594 Installing symlink pointing to librte_meter.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so 00:03:18.594 Installing symlink pointing to librte_ethdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so.23 00:03:18.594 Installing symlink pointing to librte_ethdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so 00:03:18.594 Installing symlink pointing to librte_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so.23 00:03:18.594 Installing symlink pointing to librte_pci.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so 00:03:18.594 Installing symlink pointing to librte_cmdline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so.23 00:03:18.594 Installing symlink pointing to librte_cmdline.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so 00:03:18.594 Installing symlink pointing to librte_metrics.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so.23 00:03:18.594 Installing symlink pointing to librte_metrics.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so 00:03:18.594 Installing symlink pointing to librte_hash.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so.23 00:03:18.594 Installing symlink pointing to librte_hash.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so 00:03:18.594 Installing symlink pointing to librte_timer.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so.23 00:03:18.594 Installing symlink pointing to librte_timer.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so 00:03:18.594 Installing symlink pointing to librte_acl.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so.23 00:03:18.594 Installing symlink pointing to librte_acl.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so 00:03:18.594 Installing symlink pointing to librte_bbdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so.23 00:03:18.594 Installing symlink pointing to librte_bbdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so 00:03:18.594 Installing symlink pointing to librte_bitratestats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so.23 00:03:18.594 Installing symlink pointing to librte_bitratestats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so 00:03:18.594 Installing symlink pointing to librte_bpf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so.23 00:03:18.594 Installing symlink pointing to librte_bpf.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so 00:03:18.594 Installing symlink pointing to librte_cfgfile.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so.23 00:03:18.594 Installing symlink pointing to librte_cfgfile.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so 00:03:18.594 Installing symlink pointing to librte_compressdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so.23 00:03:18.594 Installing symlink pointing to librte_compressdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so 00:03:18.594 Installing symlink pointing to librte_cryptodev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so.23 00:03:18.594 Installing symlink pointing to librte_cryptodev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so 00:03:18.594 Installing symlink pointing to librte_distributor.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so.23 00:03:18.594 Installing symlink pointing to librte_distributor.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so 00:03:18.594 Installing symlink pointing to librte_efd.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so.23 00:03:18.594 Installing symlink pointing to librte_efd.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so 00:03:18.594 Installing symlink pointing to librte_eventdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so.23 00:03:18.594 Installing symlink pointing to librte_eventdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so 00:03:18.594 Installing symlink pointing to librte_gpudev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so.23 00:03:18.594 './librte_bus_pci.so' -> 'dpdk/pmds-23.0/librte_bus_pci.so' 00:03:18.594 './librte_bus_pci.so.23' -> 'dpdk/pmds-23.0/librte_bus_pci.so.23' 00:03:18.594 './librte_bus_pci.so.23.0' -> 'dpdk/pmds-23.0/librte_bus_pci.so.23.0' 00:03:18.594 './librte_bus_vdev.so' -> 'dpdk/pmds-23.0/librte_bus_vdev.so' 00:03:18.594 './librte_bus_vdev.so.23' -> 'dpdk/pmds-23.0/librte_bus_vdev.so.23' 00:03:18.594 './librte_bus_vdev.so.23.0' -> 'dpdk/pmds-23.0/librte_bus_vdev.so.23.0' 00:03:18.594 './librte_mempool_ring.so' -> 'dpdk/pmds-23.0/librte_mempool_ring.so' 00:03:18.594 './librte_mempool_ring.so.23' -> 'dpdk/pmds-23.0/librte_mempool_ring.so.23' 00:03:18.594 './librte_mempool_ring.so.23.0' -> 'dpdk/pmds-23.0/librte_mempool_ring.so.23.0' 00:03:18.594 './librte_net_i40e.so' -> 'dpdk/pmds-23.0/librte_net_i40e.so' 00:03:18.594 './librte_net_i40e.so.23' -> 'dpdk/pmds-23.0/librte_net_i40e.so.23' 00:03:18.594 './librte_net_i40e.so.23.0' -> 'dpdk/pmds-23.0/librte_net_i40e.so.23.0' 00:03:18.594 Installing symlink pointing to librte_gpudev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so 00:03:18.594 Installing symlink pointing to librte_gro.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so.23 00:03:18.594 Installing symlink pointing to librte_gro.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so 00:03:18.594 Installing symlink pointing to librte_gso.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so.23 00:03:18.594 Installing symlink pointing to librte_gso.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so 00:03:18.594 Installing symlink pointing to librte_ip_frag.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so.23 00:03:18.594 Installing symlink pointing to librte_ip_frag.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so 00:03:18.594 Installing symlink pointing to librte_jobstats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so.23 00:03:18.594 Installing symlink pointing to librte_jobstats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so 00:03:18.594 Installing symlink pointing to librte_latencystats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so.23 00:03:18.594 Installing symlink pointing to librte_latencystats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so 00:03:18.594 Installing symlink pointing to librte_lpm.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so.23 00:03:18.594 Installing symlink pointing to librte_lpm.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so 00:03:18.594 Installing symlink pointing to librte_member.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so.23 00:03:18.594 Installing symlink pointing to librte_member.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so 00:03:18.594 Installing symlink pointing to librte_pcapng.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so.23 00:03:18.594 Installing symlink pointing to librte_pcapng.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so 00:03:18.594 Installing symlink pointing to librte_power.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so.23 00:03:18.594 Installing symlink pointing to librte_power.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so 00:03:18.594 Installing symlink pointing to librte_rawdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so.23 00:03:18.594 Installing symlink pointing to librte_rawdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so 00:03:18.594 Installing symlink pointing to librte_regexdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so.23 00:03:18.594 Installing symlink pointing to librte_regexdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so 00:03:18.594 Installing symlink pointing to librte_dmadev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so.23 00:03:18.594 Installing symlink pointing to librte_dmadev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so 00:03:18.594 Installing symlink pointing to librte_rib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so.23 00:03:18.594 Installing symlink pointing to librte_rib.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so 00:03:18.594 Installing symlink pointing to librte_reorder.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so.23 00:03:18.594 Installing symlink pointing to librte_reorder.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so 00:03:18.594 Installing symlink pointing to librte_sched.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so.23 00:03:18.594 Installing symlink pointing to librte_sched.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so 00:03:18.594 Installing symlink pointing to librte_security.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so.23 00:03:18.594 Installing symlink pointing to librte_security.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so 00:03:18.594 Installing symlink pointing to librte_stack.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so.23 00:03:18.594 Installing symlink pointing to librte_stack.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so 00:03:18.594 Installing symlink pointing to librte_vhost.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so.23 00:03:18.594 Installing symlink pointing to librte_vhost.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so 00:03:18.594 Installing symlink pointing to librte_ipsec.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so.23 00:03:18.594 Installing symlink pointing to librte_ipsec.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so 00:03:18.594 Installing symlink pointing to librte_fib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so.23 00:03:18.594 Installing symlink pointing to librte_fib.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so 00:03:18.594 Installing symlink pointing to librte_port.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so.23 00:03:18.594 Installing symlink pointing to librte_port.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so 00:03:18.594 Installing symlink pointing to librte_pdump.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so.23 00:03:18.594 Installing symlink pointing to librte_pdump.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so 00:03:18.594 Installing symlink pointing to librte_table.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so.23 00:03:18.594 Installing symlink pointing to librte_table.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so 00:03:18.594 Installing symlink pointing to librte_pipeline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so.23 00:03:18.595 Installing symlink pointing to librte_pipeline.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so 00:03:18.595 Installing symlink pointing to librte_graph.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so.23 00:03:18.595 Installing symlink pointing to librte_graph.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so 00:03:18.595 Installing symlink pointing to librte_node.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so.23 00:03:18.595 Installing symlink pointing to librte_node.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so 00:03:18.595 Installing symlink pointing to librte_bus_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so.23 00:03:18.595 Installing symlink pointing to librte_bus_pci.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so 00:03:18.595 Installing symlink pointing to librte_bus_vdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so.23 00:03:18.595 Installing symlink pointing to librte_bus_vdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so 00:03:18.595 Installing symlink pointing to librte_mempool_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so.23 00:03:18.595 Installing symlink pointing to librte_mempool_ring.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so 00:03:18.595 Installing symlink pointing to librte_net_i40e.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so.23 00:03:18.595 Installing symlink pointing to librte_net_i40e.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so 00:03:18.595 Running custom install script '/bin/sh /home/vagrant/spdk_repo/dpdk/config/../buildtools/symlink-drivers-solibs.sh lib dpdk/pmds-23.0' 00:03:18.595 18:53:35 build_native_dpdk -- common/autobuild_common.sh@220 -- $ cat 00:03:18.595 18:53:35 build_native_dpdk -- common/autobuild_common.sh@225 -- $ cd /home/vagrant/spdk_repo/spdk 00:03:18.595 00:03:18.595 real 0m43.539s 00:03:18.595 user 4m16.473s 00:03:18.595 sys 0m48.685s 00:03:18.595 ************************************ 00:03:18.595 END TEST build_native_dpdk 00:03:18.595 ************************************ 00:03:18.595 18:53:35 build_native_dpdk -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:03:18.595 18:53:35 build_native_dpdk -- common/autotest_common.sh@10 -- $ set +x 00:03:18.595 18:53:36 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:03:18.595 18:53:36 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:03:18.595 18:53:36 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:03:18.595 18:53:36 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:03:18.595 18:53:36 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:03:18.595 18:53:36 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:03:18.595 18:53:36 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:03:18.595 18:53:36 -- spdk/autobuild.sh@67 -- $ /home/vagrant/spdk_repo/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build --with-shared 00:03:18.853 Using /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig for additional libs... 00:03:18.853 DPDK libraries: /home/vagrant/spdk_repo/dpdk/build/lib 00:03:18.853 DPDK includes: //home/vagrant/spdk_repo/dpdk/build/include 00:03:18.853 Using default SPDK env in /home/vagrant/spdk_repo/spdk/lib/env_dpdk 00:03:19.112 Using 'verbs' RDMA provider 00:03:35.444 Configuring ISA-L (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal.log)...done. 00:03:53.554 Configuring ISA-L-crypto (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal-crypto.log)...done. 00:03:53.554 Creating mk/config.mk...done. 00:03:53.554 Creating mk/cc.flags.mk...done. 00:03:53.554 Type 'make' to build. 00:03:53.554 18:54:09 -- spdk/autobuild.sh@70 -- $ run_test make make -j10 00:03:53.554 18:54:09 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:03:53.554 18:54:09 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:03:53.554 18:54:09 -- common/autotest_common.sh@10 -- $ set +x 00:03:53.554 ************************************ 00:03:53.554 START TEST make 00:03:53.554 ************************************ 00:03:53.554 18:54:09 make -- common/autotest_common.sh@1129 -- $ make -j10 00:03:53.554 make[1]: Nothing to be done for 'all'. 00:04:32.328 CC lib/ut/ut.o 00:04:32.328 CC lib/ut_mock/mock.o 00:04:32.328 CC lib/log/log.o 00:04:32.328 CC lib/log/log_flags.o 00:04:32.328 CC lib/log/log_deprecated.o 00:04:32.328 LIB libspdk_ut.a 00:04:32.328 LIB libspdk_ut_mock.a 00:04:32.328 LIB libspdk_log.a 00:04:32.328 SO libspdk_ut.so.2.0 00:04:32.328 SO libspdk_ut_mock.so.6.0 00:04:32.328 SO libspdk_log.so.7.1 00:04:32.328 SYMLINK libspdk_ut.so 00:04:32.328 SYMLINK libspdk_ut_mock.so 00:04:32.328 SYMLINK libspdk_log.so 00:04:32.328 CC lib/util/bit_array.o 00:04:32.328 CC lib/util/cpuset.o 00:04:32.328 CC lib/util/base64.o 00:04:32.328 CC lib/util/crc32c.o 00:04:32.328 CC lib/util/crc32.o 00:04:32.328 CC lib/util/crc16.o 00:04:32.328 CC lib/ioat/ioat.o 00:04:32.328 CC lib/dma/dma.o 00:04:32.328 CXX lib/trace_parser/trace.o 00:04:32.328 CC lib/vfio_user/host/vfio_user_pci.o 00:04:32.328 CC lib/util/crc32_ieee.o 00:04:32.328 CC lib/util/crc64.o 00:04:32.328 CC lib/util/dif.o 00:04:32.328 CC lib/util/fd.o 00:04:32.328 CC lib/util/fd_group.o 00:04:32.328 LIB libspdk_dma.a 00:04:32.328 CC lib/util/file.o 00:04:32.328 SO libspdk_dma.so.5.0 00:04:32.328 CC lib/util/hexlify.o 00:04:32.328 CC lib/vfio_user/host/vfio_user.o 00:04:32.328 SYMLINK libspdk_dma.so 00:04:32.328 CC lib/util/iov.o 00:04:32.328 LIB libspdk_ioat.a 00:04:32.328 CC lib/util/math.o 00:04:32.328 SO libspdk_ioat.so.7.0 00:04:32.589 CC lib/util/net.o 00:04:32.589 SYMLINK libspdk_ioat.so 00:04:32.589 CC lib/util/pipe.o 00:04:32.589 CC lib/util/strerror_tls.o 00:04:32.589 CC lib/util/string.o 00:04:32.589 CC lib/util/uuid.o 00:04:32.589 LIB libspdk_vfio_user.a 00:04:32.589 CC lib/util/xor.o 00:04:32.589 CC lib/util/zipf.o 00:04:32.589 SO libspdk_vfio_user.so.5.0 00:04:32.589 CC lib/util/md5.o 00:04:32.589 SYMLINK libspdk_vfio_user.so 00:04:32.849 LIB libspdk_util.a 00:04:33.109 SO libspdk_util.so.10.1 00:04:33.109 LIB libspdk_trace_parser.a 00:04:33.109 SYMLINK libspdk_util.so 00:04:33.109 SO libspdk_trace_parser.so.6.0 00:04:33.370 SYMLINK libspdk_trace_parser.so 00:04:33.370 CC lib/conf/conf.o 00:04:33.370 CC lib/env_dpdk/env.o 00:04:33.370 CC lib/rdma_utils/rdma_utils.o 00:04:33.370 CC lib/env_dpdk/pci.o 00:04:33.370 CC lib/idxd/idxd.o 00:04:33.370 CC lib/env_dpdk/memory.o 00:04:33.370 CC lib/idxd/idxd_user.o 00:04:33.370 CC lib/env_dpdk/init.o 00:04:33.370 CC lib/vmd/vmd.o 00:04:33.370 CC lib/json/json_parse.o 00:04:33.630 LIB libspdk_conf.a 00:04:33.630 CC lib/json/json_util.o 00:04:33.630 SO libspdk_conf.so.6.0 00:04:33.630 CC lib/json/json_write.o 00:04:33.630 LIB libspdk_rdma_utils.a 00:04:33.630 SYMLINK libspdk_conf.so 00:04:33.630 CC lib/idxd/idxd_kernel.o 00:04:33.630 SO libspdk_rdma_utils.so.1.0 00:04:33.630 CC lib/env_dpdk/threads.o 00:04:33.630 SYMLINK libspdk_rdma_utils.so 00:04:33.630 CC lib/env_dpdk/pci_ioat.o 00:04:33.630 CC lib/env_dpdk/pci_virtio.o 00:04:33.630 CC lib/env_dpdk/pci_vmd.o 00:04:33.630 CC lib/vmd/led.o 00:04:33.630 CC lib/env_dpdk/pci_idxd.o 00:04:33.890 CC lib/env_dpdk/pci_event.o 00:04:33.890 CC lib/env_dpdk/sigbus_handler.o 00:04:33.890 CC lib/env_dpdk/pci_dpdk.o 00:04:33.890 LIB libspdk_json.a 00:04:33.890 SO libspdk_json.so.6.0 00:04:33.890 CC lib/env_dpdk/pci_dpdk_2207.o 00:04:33.890 CC lib/env_dpdk/pci_dpdk_2211.o 00:04:33.890 SYMLINK libspdk_json.so 00:04:33.890 CC lib/rdma_provider/common.o 00:04:33.890 CC lib/rdma_provider/rdma_provider_verbs.o 00:04:33.890 LIB libspdk_idxd.a 00:04:33.890 LIB libspdk_vmd.a 00:04:33.890 SO libspdk_idxd.so.12.1 00:04:34.150 SO libspdk_vmd.so.6.0 00:04:34.150 SYMLINK libspdk_idxd.so 00:04:34.150 CC lib/jsonrpc/jsonrpc_server.o 00:04:34.150 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:04:34.150 CC lib/jsonrpc/jsonrpc_client.o 00:04:34.150 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:04:34.150 SYMLINK libspdk_vmd.so 00:04:34.150 LIB libspdk_rdma_provider.a 00:04:34.150 SO libspdk_rdma_provider.so.7.0 00:04:34.150 SYMLINK libspdk_rdma_provider.so 00:04:34.473 LIB libspdk_jsonrpc.a 00:04:34.473 SO libspdk_jsonrpc.so.6.0 00:04:34.473 SYMLINK libspdk_jsonrpc.so 00:04:34.749 LIB libspdk_env_dpdk.a 00:04:35.010 CC lib/rpc/rpc.o 00:04:35.010 SO libspdk_env_dpdk.so.15.1 00:04:35.010 SYMLINK libspdk_env_dpdk.so 00:04:35.010 LIB libspdk_rpc.a 00:04:35.270 SO libspdk_rpc.so.6.0 00:04:35.270 SYMLINK libspdk_rpc.so 00:04:35.530 CC lib/keyring/keyring.o 00:04:35.530 CC lib/notify/notify_rpc.o 00:04:35.530 CC lib/notify/notify.o 00:04:35.530 CC lib/keyring/keyring_rpc.o 00:04:35.530 CC lib/trace/trace.o 00:04:35.530 CC lib/trace/trace_flags.o 00:04:35.530 CC lib/trace/trace_rpc.o 00:04:35.790 LIB libspdk_notify.a 00:04:35.790 SO libspdk_notify.so.6.0 00:04:35.790 LIB libspdk_keyring.a 00:04:35.790 LIB libspdk_trace.a 00:04:35.790 SYMLINK libspdk_notify.so 00:04:35.790 SO libspdk_keyring.so.2.0 00:04:35.790 SO libspdk_trace.so.11.0 00:04:36.050 SYMLINK libspdk_keyring.so 00:04:36.050 SYMLINK libspdk_trace.so 00:04:36.311 CC lib/thread/thread.o 00:04:36.311 CC lib/thread/iobuf.o 00:04:36.311 CC lib/sock/sock.o 00:04:36.311 CC lib/sock/sock_rpc.o 00:04:36.882 LIB libspdk_sock.a 00:04:36.882 SO libspdk_sock.so.10.0 00:04:36.882 SYMLINK libspdk_sock.so 00:04:37.453 CC lib/nvme/nvme_ctrlr_cmd.o 00:04:37.453 CC lib/nvme/nvme_ctrlr.o 00:04:37.453 CC lib/nvme/nvme_fabric.o 00:04:37.453 CC lib/nvme/nvme_ns_cmd.o 00:04:37.453 CC lib/nvme/nvme_ns.o 00:04:37.453 CC lib/nvme/nvme_pcie_common.o 00:04:37.453 CC lib/nvme/nvme.o 00:04:37.453 CC lib/nvme/nvme_pcie.o 00:04:37.453 CC lib/nvme/nvme_qpair.o 00:04:38.032 LIB libspdk_thread.a 00:04:38.032 CC lib/nvme/nvme_quirks.o 00:04:38.032 SO libspdk_thread.so.11.0 00:04:38.032 CC lib/nvme/nvme_transport.o 00:04:38.032 SYMLINK libspdk_thread.so 00:04:38.032 CC lib/nvme/nvme_discovery.o 00:04:38.032 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:04:38.032 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:04:38.032 CC lib/nvme/nvme_tcp.o 00:04:38.032 CC lib/nvme/nvme_opal.o 00:04:38.032 CC lib/nvme/nvme_io_msg.o 00:04:38.290 CC lib/nvme/nvme_poll_group.o 00:04:38.290 CC lib/nvme/nvme_zns.o 00:04:38.550 CC lib/nvme/nvme_stubs.o 00:04:38.550 CC lib/nvme/nvme_auth.o 00:04:38.550 CC lib/nvme/nvme_cuse.o 00:04:38.550 CC lib/nvme/nvme_rdma.o 00:04:38.809 CC lib/accel/accel.o 00:04:38.809 CC lib/blob/blobstore.o 00:04:39.068 CC lib/virtio/virtio.o 00:04:39.068 CC lib/init/json_config.o 00:04:39.068 CC lib/fsdev/fsdev.o 00:04:39.328 CC lib/init/subsystem.o 00:04:39.328 CC lib/accel/accel_rpc.o 00:04:39.328 CC lib/init/subsystem_rpc.o 00:04:39.328 CC lib/virtio/virtio_vhost_user.o 00:04:39.588 CC lib/virtio/virtio_vfio_user.o 00:04:39.588 CC lib/init/rpc.o 00:04:39.588 CC lib/fsdev/fsdev_io.o 00:04:39.588 CC lib/accel/accel_sw.o 00:04:39.588 CC lib/blob/request.o 00:04:39.588 CC lib/fsdev/fsdev_rpc.o 00:04:39.588 LIB libspdk_init.a 00:04:39.588 CC lib/blob/zeroes.o 00:04:39.847 SO libspdk_init.so.6.0 00:04:39.847 CC lib/virtio/virtio_pci.o 00:04:39.847 SYMLINK libspdk_init.so 00:04:39.847 CC lib/blob/blob_bs_dev.o 00:04:39.847 LIB libspdk_fsdev.a 00:04:39.847 LIB libspdk_accel.a 00:04:39.847 SO libspdk_fsdev.so.2.0 00:04:39.847 CC lib/event/reactor.o 00:04:39.847 CC lib/event/app.o 00:04:39.847 CC lib/event/log_rpc.o 00:04:39.847 SO libspdk_accel.so.16.0 00:04:39.847 LIB libspdk_nvme.a 00:04:40.107 CC lib/event/app_rpc.o 00:04:40.107 SYMLINK libspdk_fsdev.so 00:04:40.107 CC lib/event/scheduler_static.o 00:04:40.107 SYMLINK libspdk_accel.so 00:04:40.107 LIB libspdk_virtio.a 00:04:40.107 SO libspdk_virtio.so.7.0 00:04:40.107 SYMLINK libspdk_virtio.so 00:04:40.107 SO libspdk_nvme.so.15.0 00:04:40.107 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:04:40.107 CC lib/bdev/bdev.o 00:04:40.107 CC lib/bdev/bdev_rpc.o 00:04:40.107 CC lib/bdev/part.o 00:04:40.107 CC lib/bdev/bdev_zone.o 00:04:40.367 CC lib/bdev/scsi_nvme.o 00:04:40.367 SYMLINK libspdk_nvme.so 00:04:40.367 LIB libspdk_event.a 00:04:40.627 SO libspdk_event.so.14.0 00:04:40.627 SYMLINK libspdk_event.so 00:04:40.886 LIB libspdk_fuse_dispatcher.a 00:04:40.886 SO libspdk_fuse_dispatcher.so.1.0 00:04:40.886 SYMLINK libspdk_fuse_dispatcher.so 00:04:42.268 LIB libspdk_blob.a 00:04:42.268 SO libspdk_blob.so.12.0 00:04:42.268 SYMLINK libspdk_blob.so 00:04:42.838 CC lib/lvol/lvol.o 00:04:42.838 CC lib/blobfs/blobfs.o 00:04:42.838 CC lib/blobfs/tree.o 00:04:42.838 LIB libspdk_bdev.a 00:04:43.098 SO libspdk_bdev.so.17.0 00:04:43.098 SYMLINK libspdk_bdev.so 00:04:43.359 CC lib/nbd/nbd.o 00:04:43.359 CC lib/nbd/nbd_rpc.o 00:04:43.359 CC lib/ftl/ftl_core.o 00:04:43.359 CC lib/ftl/ftl_init.o 00:04:43.359 CC lib/scsi/dev.o 00:04:43.359 CC lib/ftl/ftl_layout.o 00:04:43.359 CC lib/ublk/ublk.o 00:04:43.359 CC lib/nvmf/ctrlr.o 00:04:43.359 CC lib/ublk/ublk_rpc.o 00:04:43.620 LIB libspdk_blobfs.a 00:04:43.620 CC lib/ftl/ftl_debug.o 00:04:43.620 SO libspdk_blobfs.so.11.0 00:04:43.620 CC lib/scsi/lun.o 00:04:43.620 SYMLINK libspdk_blobfs.so 00:04:43.620 CC lib/scsi/port.o 00:04:43.620 LIB libspdk_lvol.a 00:04:43.620 CC lib/scsi/scsi.o 00:04:43.620 CC lib/ftl/ftl_io.o 00:04:43.620 SO libspdk_lvol.so.11.0 00:04:43.620 SYMLINK libspdk_lvol.so 00:04:43.620 CC lib/ftl/ftl_sb.o 00:04:43.620 CC lib/nvmf/ctrlr_discovery.o 00:04:43.620 CC lib/nvmf/ctrlr_bdev.o 00:04:43.880 CC lib/nvmf/subsystem.o 00:04:43.880 LIB libspdk_nbd.a 00:04:43.880 CC lib/scsi/scsi_bdev.o 00:04:43.880 SO libspdk_nbd.so.7.0 00:04:43.880 SYMLINK libspdk_nbd.so 00:04:43.880 CC lib/nvmf/nvmf.o 00:04:43.880 CC lib/nvmf/nvmf_rpc.o 00:04:43.880 CC lib/nvmf/transport.o 00:04:43.880 CC lib/ftl/ftl_l2p.o 00:04:43.880 LIB libspdk_ublk.a 00:04:44.140 SO libspdk_ublk.so.3.0 00:04:44.140 SYMLINK libspdk_ublk.so 00:04:44.140 CC lib/nvmf/tcp.o 00:04:44.140 CC lib/ftl/ftl_l2p_flat.o 00:04:44.140 CC lib/nvmf/stubs.o 00:04:44.400 CC lib/scsi/scsi_pr.o 00:04:44.400 CC lib/ftl/ftl_nv_cache.o 00:04:44.400 CC lib/nvmf/mdns_server.o 00:04:44.665 CC lib/scsi/scsi_rpc.o 00:04:44.665 CC lib/nvmf/rdma.o 00:04:44.665 CC lib/nvmf/auth.o 00:04:44.665 CC lib/ftl/ftl_band.o 00:04:44.665 CC lib/scsi/task.o 00:04:44.924 CC lib/ftl/ftl_band_ops.o 00:04:44.924 CC lib/ftl/ftl_writer.o 00:04:44.924 LIB libspdk_scsi.a 00:04:45.183 SO libspdk_scsi.so.9.0 00:04:45.183 CC lib/ftl/ftl_rq.o 00:04:45.183 SYMLINK libspdk_scsi.so 00:04:45.183 CC lib/ftl/ftl_reloc.o 00:04:45.183 CC lib/ftl/ftl_l2p_cache.o 00:04:45.183 CC lib/ftl/ftl_p2l.o 00:04:45.183 CC lib/ftl/ftl_p2l_log.o 00:04:45.442 CC lib/iscsi/conn.o 00:04:45.442 CC lib/vhost/vhost.o 00:04:45.442 CC lib/vhost/vhost_rpc.o 00:04:45.442 CC lib/ftl/mngt/ftl_mngt.o 00:04:45.442 CC lib/iscsi/init_grp.o 00:04:45.442 CC lib/iscsi/iscsi.o 00:04:45.703 CC lib/iscsi/param.o 00:04:45.703 CC lib/iscsi/portal_grp.o 00:04:45.703 CC lib/iscsi/tgt_node.o 00:04:45.703 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:04:45.703 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:04:45.963 CC lib/iscsi/iscsi_subsystem.o 00:04:45.963 CC lib/ftl/mngt/ftl_mngt_startup.o 00:04:45.963 CC lib/iscsi/iscsi_rpc.o 00:04:45.963 CC lib/iscsi/task.o 00:04:45.963 CC lib/vhost/vhost_scsi.o 00:04:45.963 CC lib/vhost/vhost_blk.o 00:04:46.224 CC lib/ftl/mngt/ftl_mngt_md.o 00:04:46.224 CC lib/vhost/rte_vhost_user.o 00:04:46.224 CC lib/ftl/mngt/ftl_mngt_misc.o 00:04:46.224 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:04:46.224 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:04:46.484 CC lib/ftl/mngt/ftl_mngt_band.o 00:04:46.484 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:04:46.484 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:04:46.484 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:04:46.484 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:04:46.484 CC lib/ftl/utils/ftl_conf.o 00:04:46.484 CC lib/ftl/utils/ftl_md.o 00:04:46.745 CC lib/ftl/utils/ftl_mempool.o 00:04:46.745 CC lib/ftl/utils/ftl_bitmap.o 00:04:46.745 CC lib/ftl/utils/ftl_property.o 00:04:46.745 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:04:46.745 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:04:46.745 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:04:47.005 LIB libspdk_nvmf.a 00:04:47.005 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:04:47.005 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:04:47.005 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:04:47.005 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:04:47.005 LIB libspdk_iscsi.a 00:04:47.005 CC lib/ftl/upgrade/ftl_sb_v3.o 00:04:47.005 CC lib/ftl/upgrade/ftl_sb_v5.o 00:04:47.005 CC lib/ftl/nvc/ftl_nvc_dev.o 00:04:47.005 SO libspdk_nvmf.so.20.0 00:04:47.266 SO libspdk_iscsi.so.8.0 00:04:47.266 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:04:47.266 LIB libspdk_vhost.a 00:04:47.266 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:04:47.266 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:04:47.266 CC lib/ftl/base/ftl_base_dev.o 00:04:47.266 SO libspdk_vhost.so.8.0 00:04:47.266 CC lib/ftl/base/ftl_base_bdev.o 00:04:47.266 CC lib/ftl/ftl_trace.o 00:04:47.266 SYMLINK libspdk_iscsi.so 00:04:47.266 SYMLINK libspdk_vhost.so 00:04:47.266 SYMLINK libspdk_nvmf.so 00:04:47.562 LIB libspdk_ftl.a 00:04:47.825 SO libspdk_ftl.so.9.0 00:04:48.086 SYMLINK libspdk_ftl.so 00:04:48.345 CC module/env_dpdk/env_dpdk_rpc.o 00:04:48.345 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:04:48.345 CC module/accel/error/accel_error.o 00:04:48.345 CC module/sock/posix/posix.o 00:04:48.345 CC module/scheduler/dynamic/scheduler_dynamic.o 00:04:48.345 CC module/scheduler/gscheduler/gscheduler.o 00:04:48.345 CC module/accel/ioat/accel_ioat.o 00:04:48.345 CC module/fsdev/aio/fsdev_aio.o 00:04:48.345 CC module/keyring/file/keyring.o 00:04:48.345 CC module/blob/bdev/blob_bdev.o 00:04:48.604 LIB libspdk_env_dpdk_rpc.a 00:04:48.604 SO libspdk_env_dpdk_rpc.so.6.0 00:04:48.604 SYMLINK libspdk_env_dpdk_rpc.so 00:04:48.604 CC module/keyring/file/keyring_rpc.o 00:04:48.604 LIB libspdk_scheduler_gscheduler.a 00:04:48.604 LIB libspdk_scheduler_dpdk_governor.a 00:04:48.604 CC module/accel/ioat/accel_ioat_rpc.o 00:04:48.604 SO libspdk_scheduler_gscheduler.so.4.0 00:04:48.604 SO libspdk_scheduler_dpdk_governor.so.4.0 00:04:48.604 LIB libspdk_scheduler_dynamic.a 00:04:48.604 CC module/accel/error/accel_error_rpc.o 00:04:48.604 SYMLINK libspdk_scheduler_dpdk_governor.so 00:04:48.604 SO libspdk_scheduler_dynamic.so.4.0 00:04:48.604 SYMLINK libspdk_scheduler_gscheduler.so 00:04:48.604 LIB libspdk_keyring_file.a 00:04:48.604 LIB libspdk_blob_bdev.a 00:04:48.604 SO libspdk_keyring_file.so.2.0 00:04:48.604 LIB libspdk_accel_ioat.a 00:04:48.863 SYMLINK libspdk_scheduler_dynamic.so 00:04:48.863 CC module/fsdev/aio/fsdev_aio_rpc.o 00:04:48.863 SO libspdk_accel_ioat.so.6.0 00:04:48.863 SO libspdk_blob_bdev.so.12.0 00:04:48.863 LIB libspdk_accel_error.a 00:04:48.863 SYMLINK libspdk_keyring_file.so 00:04:48.863 CC module/fsdev/aio/linux_aio_mgr.o 00:04:48.863 CC module/keyring/linux/keyring.o 00:04:48.863 SO libspdk_accel_error.so.2.0 00:04:48.863 SYMLINK libspdk_blob_bdev.so 00:04:48.863 SYMLINK libspdk_accel_ioat.so 00:04:48.863 CC module/accel/iaa/accel_iaa.o 00:04:48.863 CC module/keyring/linux/keyring_rpc.o 00:04:48.863 CC module/accel/dsa/accel_dsa.o 00:04:48.863 SYMLINK libspdk_accel_error.so 00:04:48.863 CC module/accel/iaa/accel_iaa_rpc.o 00:04:48.863 CC module/accel/dsa/accel_dsa_rpc.o 00:04:48.863 LIB libspdk_keyring_linux.a 00:04:49.122 SO libspdk_keyring_linux.so.1.0 00:04:49.122 LIB libspdk_accel_iaa.a 00:04:49.122 SYMLINK libspdk_keyring_linux.so 00:04:49.122 SO libspdk_accel_iaa.so.3.0 00:04:49.122 CC module/bdev/delay/vbdev_delay.o 00:04:49.122 LIB libspdk_accel_dsa.a 00:04:49.122 SYMLINK libspdk_accel_iaa.so 00:04:49.122 CC module/bdev/delay/vbdev_delay_rpc.o 00:04:49.122 SO libspdk_accel_dsa.so.5.0 00:04:49.122 CC module/bdev/error/vbdev_error.o 00:04:49.122 LIB libspdk_fsdev_aio.a 00:04:49.122 CC module/bdev/lvol/vbdev_lvol.o 00:04:49.122 CC module/bdev/gpt/gpt.o 00:04:49.122 CC module/blobfs/bdev/blobfs_bdev.o 00:04:49.122 CC module/bdev/malloc/bdev_malloc.o 00:04:49.122 SO libspdk_fsdev_aio.so.1.0 00:04:49.122 SYMLINK libspdk_accel_dsa.so 00:04:49.122 LIB libspdk_sock_posix.a 00:04:49.122 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:04:49.381 SO libspdk_sock_posix.so.6.0 00:04:49.381 SYMLINK libspdk_fsdev_aio.so 00:04:49.381 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:04:49.381 CC module/bdev/error/vbdev_error_rpc.o 00:04:49.381 SYMLINK libspdk_sock_posix.so 00:04:49.381 CC module/bdev/gpt/vbdev_gpt.o 00:04:49.381 LIB libspdk_blobfs_bdev.a 00:04:49.381 SO libspdk_blobfs_bdev.so.6.0 00:04:49.381 LIB libspdk_bdev_error.a 00:04:49.381 SYMLINK libspdk_blobfs_bdev.so 00:04:49.381 LIB libspdk_bdev_delay.a 00:04:49.381 SO libspdk_bdev_error.so.6.0 00:04:49.381 CC module/bdev/null/bdev_null.o 00:04:49.381 CC module/bdev/nvme/bdev_nvme.o 00:04:49.639 SO libspdk_bdev_delay.so.6.0 00:04:49.639 SYMLINK libspdk_bdev_error.so 00:04:49.639 CC module/bdev/null/bdev_null_rpc.o 00:04:49.639 CC module/bdev/passthru/vbdev_passthru.o 00:04:49.639 SYMLINK libspdk_bdev_delay.so 00:04:49.639 CC module/bdev/malloc/bdev_malloc_rpc.o 00:04:49.639 LIB libspdk_bdev_gpt.a 00:04:49.639 CC module/bdev/raid/bdev_raid.o 00:04:49.639 SO libspdk_bdev_gpt.so.6.0 00:04:49.639 SYMLINK libspdk_bdev_gpt.so 00:04:49.639 CC module/bdev/raid/bdev_raid_rpc.o 00:04:49.639 CC module/bdev/split/vbdev_split.o 00:04:49.639 LIB libspdk_bdev_lvol.a 00:04:49.639 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:04:49.639 LIB libspdk_bdev_malloc.a 00:04:49.639 SO libspdk_bdev_lvol.so.6.0 00:04:49.639 LIB libspdk_bdev_null.a 00:04:49.639 SO libspdk_bdev_malloc.so.6.0 00:04:49.898 CC module/bdev/zone_block/vbdev_zone_block.o 00:04:49.898 SO libspdk_bdev_null.so.6.0 00:04:49.898 SYMLINK libspdk_bdev_lvol.so 00:04:49.898 SYMLINK libspdk_bdev_malloc.so 00:04:49.898 CC module/bdev/raid/bdev_raid_sb.o 00:04:49.898 SYMLINK libspdk_bdev_null.so 00:04:49.898 CC module/bdev/raid/raid0.o 00:04:49.898 CC module/bdev/raid/raid1.o 00:04:49.898 LIB libspdk_bdev_passthru.a 00:04:49.898 SO libspdk_bdev_passthru.so.6.0 00:04:49.898 CC module/bdev/nvme/bdev_nvme_rpc.o 00:04:49.898 CC module/bdev/split/vbdev_split_rpc.o 00:04:49.898 SYMLINK libspdk_bdev_passthru.so 00:04:49.898 CC module/bdev/aio/bdev_aio.o 00:04:50.157 CC module/bdev/raid/concat.o 00:04:50.157 CC module/bdev/raid/raid5f.o 00:04:50.157 LIB libspdk_bdev_split.a 00:04:50.157 CC module/bdev/nvme/nvme_rpc.o 00:04:50.157 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:04:50.157 CC module/bdev/ftl/bdev_ftl.o 00:04:50.157 SO libspdk_bdev_split.so.6.0 00:04:50.157 SYMLINK libspdk_bdev_split.so 00:04:50.417 CC module/bdev/aio/bdev_aio_rpc.o 00:04:50.417 LIB libspdk_bdev_zone_block.a 00:04:50.417 SO libspdk_bdev_zone_block.so.6.0 00:04:50.417 CC module/bdev/iscsi/bdev_iscsi.o 00:04:50.417 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:04:50.417 SYMLINK libspdk_bdev_zone_block.so 00:04:50.417 CC module/bdev/nvme/bdev_mdns_client.o 00:04:50.417 CC module/bdev/ftl/bdev_ftl_rpc.o 00:04:50.417 LIB libspdk_bdev_aio.a 00:04:50.417 CC module/bdev/virtio/bdev_virtio_scsi.o 00:04:50.417 SO libspdk_bdev_aio.so.6.0 00:04:50.417 CC module/bdev/nvme/vbdev_opal.o 00:04:50.677 SYMLINK libspdk_bdev_aio.so 00:04:50.677 CC module/bdev/nvme/vbdev_opal_rpc.o 00:04:50.677 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:04:50.677 CC module/bdev/virtio/bdev_virtio_blk.o 00:04:50.677 CC module/bdev/virtio/bdev_virtio_rpc.o 00:04:50.677 LIB libspdk_bdev_ftl.a 00:04:50.677 SO libspdk_bdev_ftl.so.6.0 00:04:50.677 LIB libspdk_bdev_raid.a 00:04:50.677 LIB libspdk_bdev_iscsi.a 00:04:50.677 SYMLINK libspdk_bdev_ftl.so 00:04:50.677 SO libspdk_bdev_raid.so.6.0 00:04:50.677 SO libspdk_bdev_iscsi.so.6.0 00:04:50.677 SYMLINK libspdk_bdev_iscsi.so 00:04:50.935 SYMLINK libspdk_bdev_raid.so 00:04:50.935 LIB libspdk_bdev_virtio.a 00:04:51.193 SO libspdk_bdev_virtio.so.6.0 00:04:51.193 SYMLINK libspdk_bdev_virtio.so 00:04:52.128 LIB libspdk_bdev_nvme.a 00:04:52.128 SO libspdk_bdev_nvme.so.7.1 00:04:52.388 SYMLINK libspdk_bdev_nvme.so 00:04:52.957 CC module/event/subsystems/scheduler/scheduler.o 00:04:52.957 CC module/event/subsystems/iobuf/iobuf.o 00:04:52.957 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:04:52.957 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:04:52.957 CC module/event/subsystems/keyring/keyring.o 00:04:52.957 CC module/event/subsystems/sock/sock.o 00:04:52.957 CC module/event/subsystems/vmd/vmd.o 00:04:52.957 CC module/event/subsystems/vmd/vmd_rpc.o 00:04:52.957 CC module/event/subsystems/fsdev/fsdev.o 00:04:52.957 LIB libspdk_event_scheduler.a 00:04:52.957 LIB libspdk_event_vhost_blk.a 00:04:52.957 LIB libspdk_event_vmd.a 00:04:52.957 LIB libspdk_event_fsdev.a 00:04:52.957 LIB libspdk_event_sock.a 00:04:52.957 LIB libspdk_event_iobuf.a 00:04:52.957 LIB libspdk_event_keyring.a 00:04:52.957 SO libspdk_event_scheduler.so.4.0 00:04:52.957 SO libspdk_event_vhost_blk.so.3.0 00:04:52.957 SO libspdk_event_fsdev.so.1.0 00:04:52.957 SO libspdk_event_vmd.so.6.0 00:04:52.957 SO libspdk_event_sock.so.5.0 00:04:52.957 SO libspdk_event_iobuf.so.3.0 00:04:52.957 SO libspdk_event_keyring.so.1.0 00:04:52.957 SYMLINK libspdk_event_scheduler.so 00:04:52.957 SYMLINK libspdk_event_fsdev.so 00:04:52.957 SYMLINK libspdk_event_vhost_blk.so 00:04:52.957 SYMLINK libspdk_event_sock.so 00:04:52.957 SYMLINK libspdk_event_vmd.so 00:04:52.957 SYMLINK libspdk_event_keyring.so 00:04:52.957 SYMLINK libspdk_event_iobuf.so 00:04:53.527 CC module/event/subsystems/accel/accel.o 00:04:53.527 LIB libspdk_event_accel.a 00:04:53.527 SO libspdk_event_accel.so.6.0 00:04:53.787 SYMLINK libspdk_event_accel.so 00:04:54.046 CC module/event/subsystems/bdev/bdev.o 00:04:54.305 LIB libspdk_event_bdev.a 00:04:54.305 SO libspdk_event_bdev.so.6.0 00:04:54.305 SYMLINK libspdk_event_bdev.so 00:04:54.564 CC module/event/subsystems/nbd/nbd.o 00:04:54.564 CC module/event/subsystems/ublk/ublk.o 00:04:54.564 CC module/event/subsystems/scsi/scsi.o 00:04:54.823 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:04:54.823 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:04:54.823 LIB libspdk_event_nbd.a 00:04:54.823 LIB libspdk_event_ublk.a 00:04:54.823 SO libspdk_event_nbd.so.6.0 00:04:54.823 LIB libspdk_event_scsi.a 00:04:54.823 SO libspdk_event_ublk.so.3.0 00:04:54.823 SYMLINK libspdk_event_nbd.so 00:04:54.823 SO libspdk_event_scsi.so.6.0 00:04:54.823 SYMLINK libspdk_event_ublk.so 00:04:54.823 SYMLINK libspdk_event_scsi.so 00:04:54.823 LIB libspdk_event_nvmf.a 00:04:55.082 SO libspdk_event_nvmf.so.6.0 00:04:55.082 SYMLINK libspdk_event_nvmf.so 00:04:55.341 CC module/event/subsystems/iscsi/iscsi.o 00:04:55.341 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:04:55.341 LIB libspdk_event_vhost_scsi.a 00:04:55.341 LIB libspdk_event_iscsi.a 00:04:55.601 SO libspdk_event_vhost_scsi.so.3.0 00:04:55.601 SO libspdk_event_iscsi.so.6.0 00:04:55.601 SYMLINK libspdk_event_iscsi.so 00:04:55.601 SYMLINK libspdk_event_vhost_scsi.so 00:04:55.861 SO libspdk.so.6.0 00:04:55.861 SYMLINK libspdk.so 00:04:56.120 CXX app/trace/trace.o 00:04:56.120 CC app/trace_record/trace_record.o 00:04:56.120 CC examples/interrupt_tgt/interrupt_tgt.o 00:04:56.120 CC app/nvmf_tgt/nvmf_main.o 00:04:56.120 CC app/iscsi_tgt/iscsi_tgt.o 00:04:56.120 CC test/thread/poller_perf/poller_perf.o 00:04:56.120 CC examples/util/zipf/zipf.o 00:04:56.120 CC app/spdk_tgt/spdk_tgt.o 00:04:56.120 CC examples/ioat/perf/perf.o 00:04:56.120 CC test/dma/test_dma/test_dma.o 00:04:56.380 LINK interrupt_tgt 00:04:56.380 LINK poller_perf 00:04:56.380 LINK nvmf_tgt 00:04:56.380 LINK zipf 00:04:56.380 LINK iscsi_tgt 00:04:56.380 LINK spdk_tgt 00:04:56.380 LINK spdk_trace_record 00:04:56.380 LINK ioat_perf 00:04:56.380 LINK spdk_trace 00:04:56.380 TEST_HEADER include/spdk/accel.h 00:04:56.380 TEST_HEADER include/spdk/accel_module.h 00:04:56.380 TEST_HEADER include/spdk/assert.h 00:04:56.640 TEST_HEADER include/spdk/barrier.h 00:04:56.640 TEST_HEADER include/spdk/base64.h 00:04:56.640 TEST_HEADER include/spdk/bdev.h 00:04:56.640 TEST_HEADER include/spdk/bdev_module.h 00:04:56.640 TEST_HEADER include/spdk/bdev_zone.h 00:04:56.640 TEST_HEADER include/spdk/bit_array.h 00:04:56.640 TEST_HEADER include/spdk/bit_pool.h 00:04:56.640 TEST_HEADER include/spdk/blob_bdev.h 00:04:56.640 TEST_HEADER include/spdk/blobfs_bdev.h 00:04:56.640 TEST_HEADER include/spdk/blobfs.h 00:04:56.640 TEST_HEADER include/spdk/blob.h 00:04:56.640 TEST_HEADER include/spdk/conf.h 00:04:56.640 TEST_HEADER include/spdk/config.h 00:04:56.640 TEST_HEADER include/spdk/cpuset.h 00:04:56.640 TEST_HEADER include/spdk/crc16.h 00:04:56.640 TEST_HEADER include/spdk/crc32.h 00:04:56.640 TEST_HEADER include/spdk/crc64.h 00:04:56.640 CC app/spdk_lspci/spdk_lspci.o 00:04:56.640 TEST_HEADER include/spdk/dif.h 00:04:56.640 TEST_HEADER include/spdk/dma.h 00:04:56.640 TEST_HEADER include/spdk/endian.h 00:04:56.640 TEST_HEADER include/spdk/env_dpdk.h 00:04:56.640 TEST_HEADER include/spdk/env.h 00:04:56.640 TEST_HEADER include/spdk/event.h 00:04:56.640 TEST_HEADER include/spdk/fd_group.h 00:04:56.640 TEST_HEADER include/spdk/fd.h 00:04:56.640 TEST_HEADER include/spdk/file.h 00:04:56.640 TEST_HEADER include/spdk/fsdev.h 00:04:56.640 TEST_HEADER include/spdk/fsdev_module.h 00:04:56.640 TEST_HEADER include/spdk/ftl.h 00:04:56.640 TEST_HEADER include/spdk/fuse_dispatcher.h 00:04:56.640 CC examples/ioat/verify/verify.o 00:04:56.640 TEST_HEADER include/spdk/gpt_spec.h 00:04:56.640 TEST_HEADER include/spdk/hexlify.h 00:04:56.640 TEST_HEADER include/spdk/histogram_data.h 00:04:56.640 CC app/spdk_nvme_perf/perf.o 00:04:56.640 TEST_HEADER include/spdk/idxd.h 00:04:56.640 TEST_HEADER include/spdk/idxd_spec.h 00:04:56.640 TEST_HEADER include/spdk/init.h 00:04:56.640 TEST_HEADER include/spdk/ioat.h 00:04:56.640 TEST_HEADER include/spdk/ioat_spec.h 00:04:56.640 TEST_HEADER include/spdk/iscsi_spec.h 00:04:56.640 TEST_HEADER include/spdk/json.h 00:04:56.640 TEST_HEADER include/spdk/jsonrpc.h 00:04:56.640 CC app/spdk_nvme_identify/identify.o 00:04:56.640 TEST_HEADER include/spdk/keyring.h 00:04:56.640 TEST_HEADER include/spdk/keyring_module.h 00:04:56.640 TEST_HEADER include/spdk/likely.h 00:04:56.640 TEST_HEADER include/spdk/log.h 00:04:56.640 CC app/spdk_nvme_discover/discovery_aer.o 00:04:56.640 TEST_HEADER include/spdk/lvol.h 00:04:56.640 TEST_HEADER include/spdk/md5.h 00:04:56.640 TEST_HEADER include/spdk/memory.h 00:04:56.640 TEST_HEADER include/spdk/mmio.h 00:04:56.640 TEST_HEADER include/spdk/nbd.h 00:04:56.640 TEST_HEADER include/spdk/net.h 00:04:56.640 TEST_HEADER include/spdk/notify.h 00:04:56.640 TEST_HEADER include/spdk/nvme.h 00:04:56.640 TEST_HEADER include/spdk/nvme_intel.h 00:04:56.640 TEST_HEADER include/spdk/nvme_ocssd.h 00:04:56.640 CC test/app/bdev_svc/bdev_svc.o 00:04:56.640 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:04:56.640 TEST_HEADER include/spdk/nvme_spec.h 00:04:56.640 CC examples/thread/thread/thread_ex.o 00:04:56.640 TEST_HEADER include/spdk/nvme_zns.h 00:04:56.640 TEST_HEADER include/spdk/nvmf_cmd.h 00:04:56.640 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:04:56.640 TEST_HEADER include/spdk/nvmf.h 00:04:56.640 TEST_HEADER include/spdk/nvmf_spec.h 00:04:56.640 TEST_HEADER include/spdk/nvmf_transport.h 00:04:56.640 TEST_HEADER include/spdk/opal.h 00:04:56.640 TEST_HEADER include/spdk/opal_spec.h 00:04:56.640 TEST_HEADER include/spdk/pci_ids.h 00:04:56.640 TEST_HEADER include/spdk/pipe.h 00:04:56.640 TEST_HEADER include/spdk/queue.h 00:04:56.640 TEST_HEADER include/spdk/reduce.h 00:04:56.640 TEST_HEADER include/spdk/rpc.h 00:04:56.640 TEST_HEADER include/spdk/scheduler.h 00:04:56.640 TEST_HEADER include/spdk/scsi.h 00:04:56.640 TEST_HEADER include/spdk/scsi_spec.h 00:04:56.640 TEST_HEADER include/spdk/sock.h 00:04:56.640 TEST_HEADER include/spdk/stdinc.h 00:04:56.640 TEST_HEADER include/spdk/string.h 00:04:56.640 TEST_HEADER include/spdk/thread.h 00:04:56.640 TEST_HEADER include/spdk/trace.h 00:04:56.640 TEST_HEADER include/spdk/trace_parser.h 00:04:56.640 TEST_HEADER include/spdk/tree.h 00:04:56.640 TEST_HEADER include/spdk/ublk.h 00:04:56.640 TEST_HEADER include/spdk/util.h 00:04:56.640 TEST_HEADER include/spdk/uuid.h 00:04:56.640 TEST_HEADER include/spdk/version.h 00:04:56.640 TEST_HEADER include/spdk/vfio_user_pci.h 00:04:56.640 TEST_HEADER include/spdk/vfio_user_spec.h 00:04:56.640 TEST_HEADER include/spdk/vhost.h 00:04:56.640 LINK test_dma 00:04:56.640 TEST_HEADER include/spdk/vmd.h 00:04:56.640 TEST_HEADER include/spdk/xor.h 00:04:56.640 TEST_HEADER include/spdk/zipf.h 00:04:56.640 CXX test/cpp_headers/accel.o 00:04:56.640 LINK spdk_lspci 00:04:56.640 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:04:56.899 LINK verify 00:04:56.899 LINK spdk_nvme_discover 00:04:56.899 LINK bdev_svc 00:04:56.899 CXX test/cpp_headers/accel_module.o 00:04:56.899 LINK thread 00:04:56.899 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:04:56.899 CC app/spdk_top/spdk_top.o 00:04:57.159 CXX test/cpp_headers/assert.o 00:04:57.159 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:04:57.159 CC app/vhost/vhost.o 00:04:57.159 CC app/spdk_dd/spdk_dd.o 00:04:57.159 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:04:57.159 CXX test/cpp_headers/barrier.o 00:04:57.159 LINK nvme_fuzz 00:04:57.159 CC examples/sock/hello_world/hello_sock.o 00:04:57.419 CXX test/cpp_headers/base64.o 00:04:57.419 LINK vhost 00:04:57.419 CXX test/cpp_headers/bdev.o 00:04:57.419 LINK spdk_nvme_perf 00:04:57.419 LINK spdk_nvme_identify 00:04:57.419 LINK spdk_dd 00:04:57.419 LINK hello_sock 00:04:57.419 CXX test/cpp_headers/bdev_module.o 00:04:57.679 LINK vhost_fuzz 00:04:57.679 CC examples/vmd/lsvmd/lsvmd.o 00:04:57.679 CC test/env/mem_callbacks/mem_callbacks.o 00:04:57.679 CXX test/cpp_headers/bdev_zone.o 00:04:57.679 CC examples/vmd/led/led.o 00:04:57.679 CC test/env/vtophys/vtophys.o 00:04:57.679 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:04:57.679 LINK lsvmd 00:04:57.938 CC examples/idxd/perf/perf.o 00:04:57.938 LINK led 00:04:57.938 LINK mem_callbacks 00:04:57.938 LINK vtophys 00:04:57.938 CXX test/cpp_headers/bit_array.o 00:04:57.938 LINK env_dpdk_post_init 00:04:57.938 CC examples/fsdev/hello_world/hello_fsdev.o 00:04:57.938 LINK spdk_top 00:04:57.938 CXX test/cpp_headers/bit_pool.o 00:04:58.198 CC examples/accel/perf/accel_perf.o 00:04:58.198 CC test/env/memory/memory_ut.o 00:04:58.198 CC test/event/event_perf/event_perf.o 00:04:58.198 CC examples/blob/hello_world/hello_blob.o 00:04:58.198 CC test/nvme/aer/aer.o 00:04:58.198 CXX test/cpp_headers/blob_bdev.o 00:04:58.198 LINK idxd_perf 00:04:58.198 LINK hello_fsdev 00:04:58.198 CC app/fio/nvme/fio_plugin.o 00:04:58.198 LINK event_perf 00:04:58.457 CXX test/cpp_headers/blobfs_bdev.o 00:04:58.457 LINK hello_blob 00:04:58.457 CC app/fio/bdev/fio_plugin.o 00:04:58.457 CC test/rpc_client/rpc_client_test.o 00:04:58.457 LINK aer 00:04:58.457 CXX test/cpp_headers/blobfs.o 00:04:58.457 CC test/event/reactor/reactor.o 00:04:58.716 LINK rpc_client_test 00:04:58.716 LINK accel_perf 00:04:58.716 CXX test/cpp_headers/blob.o 00:04:58.716 LINK reactor 00:04:58.716 CC examples/blob/cli/blobcli.o 00:04:58.716 CC test/nvme/reset/reset.o 00:04:58.716 LINK iscsi_fuzz 00:04:58.716 CXX test/cpp_headers/conf.o 00:04:58.975 LINK spdk_nvme 00:04:58.975 CC test/event/reactor_perf/reactor_perf.o 00:04:58.975 CC test/env/pci/pci_ut.o 00:04:58.975 LINK memory_ut 00:04:58.975 CXX test/cpp_headers/config.o 00:04:58.975 LINK reset 00:04:58.975 LINK spdk_bdev 00:04:58.975 CC test/accel/dif/dif.o 00:04:58.975 CXX test/cpp_headers/cpuset.o 00:04:58.975 LINK reactor_perf 00:04:58.975 CC test/app/histogram_perf/histogram_perf.o 00:04:58.975 CXX test/cpp_headers/crc16.o 00:04:59.234 CC test/blobfs/mkfs/mkfs.o 00:04:59.234 CC test/nvme/sgl/sgl.o 00:04:59.234 LINK histogram_perf 00:04:59.234 CXX test/cpp_headers/crc32.o 00:04:59.234 CC test/nvme/overhead/overhead.o 00:04:59.234 CC test/nvme/e2edp/nvme_dp.o 00:04:59.234 LINK blobcli 00:04:59.234 CC test/event/app_repeat/app_repeat.o 00:04:59.234 LINK mkfs 00:04:59.234 LINK pci_ut 00:04:59.234 CXX test/cpp_headers/crc64.o 00:04:59.492 CC test/app/jsoncat/jsoncat.o 00:04:59.492 LINK app_repeat 00:04:59.492 LINK sgl 00:04:59.492 LINK nvme_dp 00:04:59.492 LINK overhead 00:04:59.492 CXX test/cpp_headers/dif.o 00:04:59.492 LINK jsoncat 00:04:59.492 CXX test/cpp_headers/dma.o 00:04:59.492 CC examples/nvme/hello_world/hello_world.o 00:04:59.492 CXX test/cpp_headers/endian.o 00:04:59.750 CXX test/cpp_headers/env_dpdk.o 00:04:59.750 LINK dif 00:04:59.750 CC examples/bdev/hello_world/hello_bdev.o 00:04:59.750 CC test/event/scheduler/scheduler.o 00:04:59.750 CC test/nvme/err_injection/err_injection.o 00:04:59.750 CC test/app/stub/stub.o 00:04:59.750 CC examples/bdev/bdevperf/bdevperf.o 00:04:59.750 CXX test/cpp_headers/env.o 00:04:59.750 CXX test/cpp_headers/event.o 00:04:59.750 LINK hello_world 00:04:59.750 CXX test/cpp_headers/fd_group.o 00:05:00.010 CC test/lvol/esnap/esnap.o 00:05:00.010 LINK hello_bdev 00:05:00.010 LINK err_injection 00:05:00.010 LINK scheduler 00:05:00.010 LINK stub 00:05:00.010 CC test/nvme/startup/startup.o 00:05:00.010 CXX test/cpp_headers/fd.o 00:05:00.010 CC examples/nvme/reconnect/reconnect.o 00:05:00.010 CC examples/nvme/nvme_manage/nvme_manage.o 00:05:00.010 CC examples/nvme/arbitration/arbitration.o 00:05:00.269 LINK startup 00:05:00.269 CXX test/cpp_headers/file.o 00:05:00.269 CC examples/nvme/hotplug/hotplug.o 00:05:00.269 CC examples/nvme/cmb_copy/cmb_copy.o 00:05:00.269 CC examples/nvme/abort/abort.o 00:05:00.269 CXX test/cpp_headers/fsdev.o 00:05:00.269 LINK cmb_copy 00:05:00.269 LINK reconnect 00:05:00.541 LINK hotplug 00:05:00.541 CC test/nvme/reserve/reserve.o 00:05:00.541 CXX test/cpp_headers/fsdev_module.o 00:05:00.541 LINK arbitration 00:05:00.541 CXX test/cpp_headers/ftl.o 00:05:00.541 LINK nvme_manage 00:05:00.541 CXX test/cpp_headers/fuse_dispatcher.o 00:05:00.541 LINK bdevperf 00:05:00.541 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:05:00.541 LINK reserve 00:05:00.816 LINK abort 00:05:00.816 CXX test/cpp_headers/gpt_spec.o 00:05:00.816 CXX test/cpp_headers/hexlify.o 00:05:00.816 CXX test/cpp_headers/histogram_data.o 00:05:00.816 CC test/nvme/simple_copy/simple_copy.o 00:05:00.816 CXX test/cpp_headers/idxd.o 00:05:00.816 LINK pmr_persistence 00:05:00.816 CC test/bdev/bdevio/bdevio.o 00:05:00.816 CXX test/cpp_headers/idxd_spec.o 00:05:00.816 CXX test/cpp_headers/init.o 00:05:00.816 CXX test/cpp_headers/ioat.o 00:05:00.816 CC test/nvme/boot_partition/boot_partition.o 00:05:00.816 CC test/nvme/connect_stress/connect_stress.o 00:05:00.816 CXX test/cpp_headers/ioat_spec.o 00:05:01.076 LINK simple_copy 00:05:01.076 CXX test/cpp_headers/iscsi_spec.o 00:05:01.076 LINK boot_partition 00:05:01.076 CC test/nvme/compliance/nvme_compliance.o 00:05:01.076 CC test/nvme/fused_ordering/fused_ordering.o 00:05:01.076 CXX test/cpp_headers/json.o 00:05:01.076 LINK connect_stress 00:05:01.076 CC examples/nvmf/nvmf/nvmf.o 00:05:01.076 CC test/nvme/doorbell_aers/doorbell_aers.o 00:05:01.076 LINK bdevio 00:05:01.336 CC test/nvme/fdp/fdp.o 00:05:01.336 CXX test/cpp_headers/jsonrpc.o 00:05:01.336 CXX test/cpp_headers/keyring.o 00:05:01.336 LINK fused_ordering 00:05:01.336 CC test/nvme/cuse/cuse.o 00:05:01.336 LINK doorbell_aers 00:05:01.336 CXX test/cpp_headers/keyring_module.o 00:05:01.336 CXX test/cpp_headers/likely.o 00:05:01.336 LINK nvme_compliance 00:05:01.336 CXX test/cpp_headers/log.o 00:05:01.336 CXX test/cpp_headers/lvol.o 00:05:01.336 LINK nvmf 00:05:01.596 CXX test/cpp_headers/md5.o 00:05:01.596 CXX test/cpp_headers/memory.o 00:05:01.596 CXX test/cpp_headers/mmio.o 00:05:01.596 CXX test/cpp_headers/nbd.o 00:05:01.596 CXX test/cpp_headers/net.o 00:05:01.596 LINK fdp 00:05:01.596 CXX test/cpp_headers/notify.o 00:05:01.596 CXX test/cpp_headers/nvme.o 00:05:01.596 CXX test/cpp_headers/nvme_intel.o 00:05:01.596 CXX test/cpp_headers/nvme_ocssd.o 00:05:01.596 CXX test/cpp_headers/nvme_ocssd_spec.o 00:05:01.596 CXX test/cpp_headers/nvme_spec.o 00:05:01.596 CXX test/cpp_headers/nvme_zns.o 00:05:01.596 CXX test/cpp_headers/nvmf_cmd.o 00:05:01.596 CXX test/cpp_headers/nvmf_fc_spec.o 00:05:01.596 CXX test/cpp_headers/nvmf.o 00:05:01.855 CXX test/cpp_headers/nvmf_spec.o 00:05:01.855 CXX test/cpp_headers/nvmf_transport.o 00:05:01.855 CXX test/cpp_headers/opal.o 00:05:01.855 CXX test/cpp_headers/opal_spec.o 00:05:01.855 CXX test/cpp_headers/pci_ids.o 00:05:01.855 CXX test/cpp_headers/pipe.o 00:05:01.855 CXX test/cpp_headers/queue.o 00:05:01.855 CXX test/cpp_headers/reduce.o 00:05:01.855 CXX test/cpp_headers/rpc.o 00:05:01.855 CXX test/cpp_headers/scheduler.o 00:05:01.855 CXX test/cpp_headers/scsi.o 00:05:01.855 CXX test/cpp_headers/scsi_spec.o 00:05:02.114 CXX test/cpp_headers/sock.o 00:05:02.114 CXX test/cpp_headers/stdinc.o 00:05:02.114 CXX test/cpp_headers/string.o 00:05:02.114 CXX test/cpp_headers/thread.o 00:05:02.114 CXX test/cpp_headers/trace.o 00:05:02.114 CXX test/cpp_headers/trace_parser.o 00:05:02.114 CXX test/cpp_headers/tree.o 00:05:02.114 CXX test/cpp_headers/ublk.o 00:05:02.114 CXX test/cpp_headers/util.o 00:05:02.114 CXX test/cpp_headers/uuid.o 00:05:02.114 CXX test/cpp_headers/version.o 00:05:02.114 CXX test/cpp_headers/vfio_user_pci.o 00:05:02.114 CXX test/cpp_headers/vfio_user_spec.o 00:05:02.114 CXX test/cpp_headers/vhost.o 00:05:02.114 CXX test/cpp_headers/vmd.o 00:05:02.114 CXX test/cpp_headers/xor.o 00:05:02.374 CXX test/cpp_headers/zipf.o 00:05:02.633 LINK cuse 00:05:05.174 LINK esnap 00:05:05.434 00:05:05.434 real 1m13.803s 00:05:05.434 user 5m36.446s 00:05:05.434 sys 1m5.737s 00:05:05.434 18:55:22 make -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:05:05.434 18:55:22 make -- common/autotest_common.sh@10 -- $ set +x 00:05:05.434 ************************************ 00:05:05.434 END TEST make 00:05:05.434 ************************************ 00:05:05.434 18:55:22 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:05:05.434 18:55:22 -- pm/common@29 -- $ signal_monitor_resources TERM 00:05:05.434 18:55:22 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:05:05.434 18:55:22 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:05.434 18:55:22 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:05:05.434 18:55:22 -- pm/common@44 -- $ pid=6201 00:05:05.434 18:55:22 -- pm/common@50 -- $ kill -TERM 6201 00:05:05.434 18:55:22 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:05.434 18:55:22 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:05:05.434 18:55:22 -- pm/common@44 -- $ pid=6203 00:05:05.434 18:55:22 -- pm/common@50 -- $ kill -TERM 6203 00:05:05.434 18:55:22 -- spdk/autorun.sh@26 -- $ (( SPDK_TEST_UNITTEST == 1 || SPDK_RUN_FUNCTIONAL_TEST == 1 )) 00:05:05.434 18:55:22 -- spdk/autorun.sh@27 -- $ sudo -E /home/vagrant/spdk_repo/spdk/autotest.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:05:05.694 18:55:23 -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:05:05.694 18:55:23 -- common/autotest_common.sh@1711 -- # lcov --version 00:05:05.694 18:55:23 -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:05:05.695 18:55:23 -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:05:05.695 18:55:23 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:05.695 18:55:23 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:05.695 18:55:23 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:05.695 18:55:23 -- scripts/common.sh@336 -- # IFS=.-: 00:05:05.695 18:55:23 -- scripts/common.sh@336 -- # read -ra ver1 00:05:05.695 18:55:23 -- scripts/common.sh@337 -- # IFS=.-: 00:05:05.695 18:55:23 -- scripts/common.sh@337 -- # read -ra ver2 00:05:05.695 18:55:23 -- scripts/common.sh@338 -- # local 'op=<' 00:05:05.695 18:55:23 -- scripts/common.sh@340 -- # ver1_l=2 00:05:05.695 18:55:23 -- scripts/common.sh@341 -- # ver2_l=1 00:05:05.695 18:55:23 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:05.695 18:55:23 -- scripts/common.sh@344 -- # case "$op" in 00:05:05.695 18:55:23 -- scripts/common.sh@345 -- # : 1 00:05:05.695 18:55:23 -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:05.695 18:55:23 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:05.695 18:55:23 -- scripts/common.sh@365 -- # decimal 1 00:05:05.695 18:55:23 -- scripts/common.sh@353 -- # local d=1 00:05:05.695 18:55:23 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:05.695 18:55:23 -- scripts/common.sh@355 -- # echo 1 00:05:05.695 18:55:23 -- scripts/common.sh@365 -- # ver1[v]=1 00:05:05.695 18:55:23 -- scripts/common.sh@366 -- # decimal 2 00:05:05.695 18:55:23 -- scripts/common.sh@353 -- # local d=2 00:05:05.695 18:55:23 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:05.695 18:55:23 -- scripts/common.sh@355 -- # echo 2 00:05:05.695 18:55:23 -- scripts/common.sh@366 -- # ver2[v]=2 00:05:05.695 18:55:23 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:05.695 18:55:23 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:05.695 18:55:23 -- scripts/common.sh@368 -- # return 0 00:05:05.695 18:55:23 -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:05.695 18:55:23 -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:05:05.695 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:05.695 --rc genhtml_branch_coverage=1 00:05:05.695 --rc genhtml_function_coverage=1 00:05:05.695 --rc genhtml_legend=1 00:05:05.695 --rc geninfo_all_blocks=1 00:05:05.695 --rc geninfo_unexecuted_blocks=1 00:05:05.695 00:05:05.695 ' 00:05:05.695 18:55:23 -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:05:05.695 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:05.695 --rc genhtml_branch_coverage=1 00:05:05.695 --rc genhtml_function_coverage=1 00:05:05.695 --rc genhtml_legend=1 00:05:05.695 --rc geninfo_all_blocks=1 00:05:05.695 --rc geninfo_unexecuted_blocks=1 00:05:05.695 00:05:05.695 ' 00:05:05.695 18:55:23 -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:05:05.695 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:05.695 --rc genhtml_branch_coverage=1 00:05:05.695 --rc genhtml_function_coverage=1 00:05:05.695 --rc genhtml_legend=1 00:05:05.695 --rc geninfo_all_blocks=1 00:05:05.695 --rc geninfo_unexecuted_blocks=1 00:05:05.695 00:05:05.695 ' 00:05:05.695 18:55:23 -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:05:05.695 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:05.695 --rc genhtml_branch_coverage=1 00:05:05.695 --rc genhtml_function_coverage=1 00:05:05.695 --rc genhtml_legend=1 00:05:05.695 --rc geninfo_all_blocks=1 00:05:05.695 --rc geninfo_unexecuted_blocks=1 00:05:05.695 00:05:05.695 ' 00:05:05.695 18:55:23 -- spdk/autotest.sh@25 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:05:05.695 18:55:23 -- nvmf/common.sh@7 -- # uname -s 00:05:05.695 18:55:23 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:05.695 18:55:23 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:05.695 18:55:23 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:05.695 18:55:23 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:05.695 18:55:23 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:05.695 18:55:23 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:05.695 18:55:23 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:05.695 18:55:23 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:05.695 18:55:23 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:05.695 18:55:23 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:05.695 18:55:23 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:4da6608a-058c-4cf5-8acb-b90b56f362fe 00:05:05.695 18:55:23 -- nvmf/common.sh@18 -- # NVME_HOSTID=4da6608a-058c-4cf5-8acb-b90b56f362fe 00:05:05.695 18:55:23 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:05.695 18:55:23 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:05.695 18:55:23 -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:05:05.695 18:55:23 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:05.695 18:55:23 -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:05:05.695 18:55:23 -- scripts/common.sh@15 -- # shopt -s extglob 00:05:05.695 18:55:23 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:05.695 18:55:23 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:05.695 18:55:23 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:05.695 18:55:23 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:05.695 18:55:23 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:05.695 18:55:23 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:05.695 18:55:23 -- paths/export.sh@5 -- # export PATH 00:05:05.695 18:55:23 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:05.695 18:55:23 -- nvmf/common.sh@51 -- # : 0 00:05:05.695 18:55:23 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:05.695 18:55:23 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:05.695 18:55:23 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:05.695 18:55:23 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:05.695 18:55:23 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:05.695 18:55:23 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:05.695 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:05.695 18:55:23 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:05.695 18:55:23 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:05.695 18:55:23 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:05.695 18:55:23 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:05:05.695 18:55:23 -- spdk/autotest.sh@32 -- # uname -s 00:05:05.695 18:55:23 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:05:05.695 18:55:23 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:05:05.695 18:55:23 -- spdk/autotest.sh@34 -- # mkdir -p /home/vagrant/spdk_repo/spdk/../output/coredumps 00:05:05.695 18:55:23 -- spdk/autotest.sh@39 -- # echo '|/home/vagrant/spdk_repo/spdk/scripts/core-collector.sh %P %s %t' 00:05:05.695 18:55:23 -- spdk/autotest.sh@40 -- # echo /home/vagrant/spdk_repo/spdk/../output/coredumps 00:05:05.695 18:55:23 -- spdk/autotest.sh@44 -- # modprobe nbd 00:05:05.955 18:55:23 -- spdk/autotest.sh@46 -- # type -P udevadm 00:05:05.955 18:55:23 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:05:05.955 18:55:23 -- spdk/autotest.sh@48 -- # udevadm_pid=66470 00:05:05.955 18:55:23 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:05:05.955 18:55:23 -- pm/common@17 -- # local monitor 00:05:05.955 18:55:23 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:05.955 18:55:23 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:05.955 18:55:23 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:05:05.955 18:55:23 -- pm/common@21 -- # date +%s 00:05:05.955 18:55:23 -- pm/common@25 -- # sleep 1 00:05:05.955 18:55:23 -- pm/common@21 -- # date +%s 00:05:05.955 18:55:23 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1733424923 00:05:05.955 18:55:23 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1733424923 00:05:05.955 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1733424923_collect-vmstat.pm.log 00:05:05.955 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1733424923_collect-cpu-load.pm.log 00:05:06.896 18:55:24 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:05:06.896 18:55:24 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:05:06.896 18:55:24 -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:06.896 18:55:24 -- common/autotest_common.sh@10 -- # set +x 00:05:06.896 18:55:24 -- spdk/autotest.sh@59 -- # create_test_list 00:05:06.896 18:55:24 -- common/autotest_common.sh@752 -- # xtrace_disable 00:05:06.896 18:55:24 -- common/autotest_common.sh@10 -- # set +x 00:05:06.896 18:55:24 -- spdk/autotest.sh@61 -- # dirname /home/vagrant/spdk_repo/spdk/autotest.sh 00:05:06.896 18:55:24 -- spdk/autotest.sh@61 -- # readlink -f /home/vagrant/spdk_repo/spdk 00:05:06.896 18:55:24 -- spdk/autotest.sh@61 -- # src=/home/vagrant/spdk_repo/spdk 00:05:06.896 18:55:24 -- spdk/autotest.sh@62 -- # out=/home/vagrant/spdk_repo/spdk/../output 00:05:06.896 18:55:24 -- spdk/autotest.sh@63 -- # cd /home/vagrant/spdk_repo/spdk 00:05:06.896 18:55:24 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:05:06.896 18:55:24 -- common/autotest_common.sh@1457 -- # uname 00:05:06.896 18:55:24 -- common/autotest_common.sh@1457 -- # '[' Linux = FreeBSD ']' 00:05:06.896 18:55:24 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:05:06.896 18:55:24 -- common/autotest_common.sh@1477 -- # uname 00:05:06.896 18:55:24 -- common/autotest_common.sh@1477 -- # [[ Linux = FreeBSD ]] 00:05:06.896 18:55:24 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:05:06.896 18:55:24 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:05:06.896 lcov: LCOV version 1.15 00:05:06.896 18:55:24 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /home/vagrant/spdk_repo/spdk -o /home/vagrant/spdk_repo/spdk/../output/cov_base.info 00:05:21.793 /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:05:21.793 geninfo: WARNING: GCOV did not produce any data for /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno 00:05:34.014 18:55:50 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:05:34.014 18:55:50 -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:34.014 18:55:50 -- common/autotest_common.sh@10 -- # set +x 00:05:34.014 18:55:50 -- spdk/autotest.sh@78 -- # rm -f 00:05:34.014 18:55:50 -- spdk/autotest.sh@81 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:05:34.014 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:34.014 0000:00:11.0 (1b36 0010): Already using the nvme driver 00:05:34.014 0000:00:10.0 (1b36 0010): Already using the nvme driver 00:05:34.014 18:55:51 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:05:34.014 18:55:51 -- common/autotest_common.sh@1657 -- # zoned_devs=() 00:05:34.014 18:55:51 -- common/autotest_common.sh@1657 -- # local -gA zoned_devs 00:05:34.014 18:55:51 -- common/autotest_common.sh@1658 -- # zoned_ctrls=() 00:05:34.014 18:55:51 -- common/autotest_common.sh@1658 -- # local -A zoned_ctrls 00:05:34.014 18:55:51 -- common/autotest_common.sh@1659 -- # local nvme bdf ns 00:05:34.014 18:55:51 -- common/autotest_common.sh@1668 -- # for nvme in /sys/class/nvme/nvme* 00:05:34.014 18:55:51 -- common/autotest_common.sh@1669 -- # bdf=0000:00:11.0 00:05:34.014 18:55:51 -- common/autotest_common.sh@1670 -- # for ns in "$nvme/"nvme*n* 00:05:34.014 18:55:51 -- common/autotest_common.sh@1671 -- # is_block_zoned nvme0n1 00:05:34.014 18:55:51 -- common/autotest_common.sh@1650 -- # local device=nvme0n1 00:05:34.014 18:55:51 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:05:34.014 18:55:51 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:05:34.014 18:55:51 -- common/autotest_common.sh@1670 -- # for ns in "$nvme/"nvme*n* 00:05:34.275 18:55:51 -- common/autotest_common.sh@1671 -- # is_block_zoned nvme0n2 00:05:34.275 18:55:51 -- common/autotest_common.sh@1650 -- # local device=nvme0n2 00:05:34.275 18:55:51 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n2/queue/zoned ]] 00:05:34.275 18:55:51 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:05:34.275 18:55:51 -- common/autotest_common.sh@1670 -- # for ns in "$nvme/"nvme*n* 00:05:34.275 18:55:51 -- common/autotest_common.sh@1671 -- # is_block_zoned nvme0n3 00:05:34.275 18:55:51 -- common/autotest_common.sh@1650 -- # local device=nvme0n3 00:05:34.275 18:55:51 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n3/queue/zoned ]] 00:05:34.275 18:55:51 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:05:34.275 18:55:51 -- common/autotest_common.sh@1668 -- # for nvme in /sys/class/nvme/nvme* 00:05:34.275 18:55:51 -- common/autotest_common.sh@1669 -- # bdf=0000:00:10.0 00:05:34.275 18:55:51 -- common/autotest_common.sh@1670 -- # for ns in "$nvme/"nvme*n* 00:05:34.275 18:55:51 -- common/autotest_common.sh@1671 -- # is_block_zoned nvme1n1 00:05:34.275 18:55:51 -- common/autotest_common.sh@1650 -- # local device=nvme1n1 00:05:34.275 18:55:51 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme1n1/queue/zoned ]] 00:05:34.275 18:55:51 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:05:34.275 18:55:51 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:05:34.275 18:55:51 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:34.275 18:55:51 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:34.275 18:55:51 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:05:34.275 18:55:51 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:05:34.275 18:55:51 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:05:34.275 No valid GPT data, bailing 00:05:34.275 18:55:51 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:05:34.275 18:55:51 -- scripts/common.sh@394 -- # pt= 00:05:34.275 18:55:51 -- scripts/common.sh@395 -- # return 1 00:05:34.275 18:55:51 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:05:34.275 1+0 records in 00:05:34.275 1+0 records out 00:05:34.275 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00689506 s, 152 MB/s 00:05:34.275 18:55:51 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:34.275 18:55:51 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:34.275 18:55:51 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n2 00:05:34.275 18:55:51 -- scripts/common.sh@381 -- # local block=/dev/nvme0n2 pt 00:05:34.275 18:55:51 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n2 00:05:34.275 No valid GPT data, bailing 00:05:34.275 18:55:51 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n2 00:05:34.275 18:55:51 -- scripts/common.sh@394 -- # pt= 00:05:34.275 18:55:51 -- scripts/common.sh@395 -- # return 1 00:05:34.275 18:55:51 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n2 bs=1M count=1 00:05:34.275 1+0 records in 00:05:34.275 1+0 records out 00:05:34.275 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00660438 s, 159 MB/s 00:05:34.275 18:55:51 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:34.275 18:55:51 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:34.275 18:55:51 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n3 00:05:34.275 18:55:51 -- scripts/common.sh@381 -- # local block=/dev/nvme0n3 pt 00:05:34.275 18:55:51 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n3 00:05:34.275 No valid GPT data, bailing 00:05:34.275 18:55:51 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n3 00:05:34.275 18:55:51 -- scripts/common.sh@394 -- # pt= 00:05:34.275 18:55:51 -- scripts/common.sh@395 -- # return 1 00:05:34.275 18:55:51 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n3 bs=1M count=1 00:05:34.275 1+0 records in 00:05:34.275 1+0 records out 00:05:34.275 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00443851 s, 236 MB/s 00:05:34.275 18:55:51 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:34.275 18:55:51 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:34.275 18:55:51 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n1 00:05:34.275 18:55:51 -- scripts/common.sh@381 -- # local block=/dev/nvme1n1 pt 00:05:34.275 18:55:51 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n1 00:05:34.536 No valid GPT data, bailing 00:05:34.536 18:55:51 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n1 00:05:34.536 18:55:51 -- scripts/common.sh@394 -- # pt= 00:05:34.536 18:55:51 -- scripts/common.sh@395 -- # return 1 00:05:34.536 18:55:51 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n1 bs=1M count=1 00:05:34.536 1+0 records in 00:05:34.536 1+0 records out 00:05:34.536 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00423834 s, 247 MB/s 00:05:34.536 18:55:51 -- spdk/autotest.sh@105 -- # sync 00:05:35.109 18:55:52 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:05:35.109 18:55:52 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:05:35.109 18:55:52 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:05:38.417 18:55:55 -- spdk/autotest.sh@111 -- # uname -s 00:05:38.417 18:55:55 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:05:38.417 18:55:55 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:05:38.417 18:55:55 -- spdk/autotest.sh@115 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:05:38.986 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:38.986 Hugepages 00:05:38.986 node hugesize free / total 00:05:38.986 node0 1048576kB 0 / 0 00:05:38.986 node0 2048kB 0 / 0 00:05:38.986 00:05:38.986 Type BDF Vendor Device NUMA Driver Device Block devices 00:05:38.986 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:05:38.986 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme1 nvme1n1 00:05:39.244 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme0 nvme0n1 nvme0n2 nvme0n3 00:05:39.244 18:55:56 -- spdk/autotest.sh@117 -- # uname -s 00:05:39.244 18:55:56 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:05:39.244 18:55:56 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:05:39.244 18:55:56 -- common/autotest_common.sh@1516 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:05:40.182 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:40.182 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:05:40.182 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:05:40.182 18:55:57 -- common/autotest_common.sh@1517 -- # sleep 1 00:05:41.563 18:55:58 -- common/autotest_common.sh@1518 -- # bdfs=() 00:05:41.563 18:55:58 -- common/autotest_common.sh@1518 -- # local bdfs 00:05:41.563 18:55:58 -- common/autotest_common.sh@1520 -- # bdfs=($(get_nvme_bdfs)) 00:05:41.563 18:55:58 -- common/autotest_common.sh@1520 -- # get_nvme_bdfs 00:05:41.563 18:55:58 -- common/autotest_common.sh@1498 -- # bdfs=() 00:05:41.563 18:55:58 -- common/autotest_common.sh@1498 -- # local bdfs 00:05:41.563 18:55:58 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:05:41.563 18:55:58 -- common/autotest_common.sh@1499 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:05:41.563 18:55:58 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:05:41.563 18:55:58 -- common/autotest_common.sh@1500 -- # (( 2 == 0 )) 00:05:41.563 18:55:58 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:05:41.563 18:55:58 -- common/autotest_common.sh@1522 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:05:41.821 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:41.821 Waiting for block devices as requested 00:05:41.821 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:05:42.080 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:05:42.080 18:55:59 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:05:42.080 18:55:59 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:00:10.0 00:05:42.080 18:55:59 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:05:42.080 18:55:59 -- common/autotest_common.sh@1487 -- # grep 0000:00:10.0/nvme/nvme 00:05:42.080 18:55:59 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:05:42.080 18:55:59 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 ]] 00:05:42.080 18:55:59 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:05:42.080 18:55:59 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme1 00:05:42.080 18:55:59 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme1 00:05:42.080 18:55:59 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme1 ]] 00:05:42.080 18:55:59 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme1 00:05:42.080 18:55:59 -- common/autotest_common.sh@1531 -- # grep oacs 00:05:42.080 18:55:59 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:05:42.080 18:55:59 -- common/autotest_common.sh@1531 -- # oacs=' 0x12a' 00:05:42.080 18:55:59 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:05:42.080 18:55:59 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:05:42.080 18:55:59 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme1 00:05:42.080 18:55:59 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:05:42.080 18:55:59 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:05:42.080 18:55:59 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:05:42.080 18:55:59 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:05:42.080 18:55:59 -- common/autotest_common.sh@1543 -- # continue 00:05:42.080 18:55:59 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:05:42.080 18:55:59 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:00:11.0 00:05:42.080 18:55:59 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:05:42.080 18:55:59 -- common/autotest_common.sh@1487 -- # grep 0000:00:11.0/nvme/nvme 00:05:42.080 18:55:59 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:05:42.080 18:55:59 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 ]] 00:05:42.080 18:55:59 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:05:42.080 18:55:59 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme0 00:05:42.080 18:55:59 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme0 00:05:42.080 18:55:59 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme0 ]] 00:05:42.080 18:55:59 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme0 00:05:42.080 18:55:59 -- common/autotest_common.sh@1531 -- # grep oacs 00:05:42.080 18:55:59 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:05:42.080 18:55:59 -- common/autotest_common.sh@1531 -- # oacs=' 0x12a' 00:05:42.080 18:55:59 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:05:42.080 18:55:59 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:05:42.080 18:55:59 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme0 00:05:42.080 18:55:59 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:05:42.080 18:55:59 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:05:42.080 18:55:59 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:05:42.080 18:55:59 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:05:42.080 18:55:59 -- common/autotest_common.sh@1543 -- # continue 00:05:42.080 18:55:59 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:05:42.080 18:55:59 -- common/autotest_common.sh@732 -- # xtrace_disable 00:05:42.081 18:55:59 -- common/autotest_common.sh@10 -- # set +x 00:05:42.340 18:55:59 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:05:42.340 18:55:59 -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:42.340 18:55:59 -- common/autotest_common.sh@10 -- # set +x 00:05:42.340 18:55:59 -- spdk/autotest.sh@126 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:05:43.278 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:43.278 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:05:43.278 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:05:43.278 18:56:00 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:05:43.278 18:56:00 -- common/autotest_common.sh@732 -- # xtrace_disable 00:05:43.278 18:56:00 -- common/autotest_common.sh@10 -- # set +x 00:05:43.278 18:56:00 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:05:43.278 18:56:00 -- common/autotest_common.sh@1578 -- # mapfile -t bdfs 00:05:43.278 18:56:00 -- common/autotest_common.sh@1578 -- # get_nvme_bdfs_by_id 0x0a54 00:05:43.278 18:56:00 -- common/autotest_common.sh@1563 -- # bdfs=() 00:05:43.278 18:56:00 -- common/autotest_common.sh@1563 -- # _bdfs=() 00:05:43.278 18:56:00 -- common/autotest_common.sh@1563 -- # local bdfs _bdfs 00:05:43.278 18:56:00 -- common/autotest_common.sh@1564 -- # _bdfs=($(get_nvme_bdfs)) 00:05:43.278 18:56:00 -- common/autotest_common.sh@1564 -- # get_nvme_bdfs 00:05:43.278 18:56:00 -- common/autotest_common.sh@1498 -- # bdfs=() 00:05:43.278 18:56:00 -- common/autotest_common.sh@1498 -- # local bdfs 00:05:43.278 18:56:00 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:05:43.278 18:56:00 -- common/autotest_common.sh@1499 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:05:43.278 18:56:00 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:05:43.537 18:56:00 -- common/autotest_common.sh@1500 -- # (( 2 == 0 )) 00:05:43.537 18:56:00 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:05:43.537 18:56:00 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:05:43.537 18:56:00 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:00:10.0/device 00:05:43.537 18:56:00 -- common/autotest_common.sh@1566 -- # device=0x0010 00:05:43.537 18:56:00 -- common/autotest_common.sh@1567 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:05:43.537 18:56:00 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:05:43.537 18:56:00 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:00:11.0/device 00:05:43.537 18:56:00 -- common/autotest_common.sh@1566 -- # device=0x0010 00:05:43.537 18:56:00 -- common/autotest_common.sh@1567 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:05:43.537 18:56:00 -- common/autotest_common.sh@1572 -- # (( 0 > 0 )) 00:05:43.537 18:56:00 -- common/autotest_common.sh@1572 -- # return 0 00:05:43.537 18:56:00 -- common/autotest_common.sh@1579 -- # [[ -z '' ]] 00:05:43.537 18:56:00 -- common/autotest_common.sh@1580 -- # return 0 00:05:43.537 18:56:00 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:05:43.537 18:56:00 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:05:43.537 18:56:00 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:05:43.537 18:56:00 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:05:43.537 18:56:00 -- spdk/autotest.sh@149 -- # timing_enter lib 00:05:43.537 18:56:00 -- common/autotest_common.sh@726 -- # xtrace_disable 00:05:43.537 18:56:00 -- common/autotest_common.sh@10 -- # set +x 00:05:43.537 18:56:00 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:05:43.537 18:56:00 -- spdk/autotest.sh@155 -- # run_test env /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:05:43.537 18:56:00 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:43.537 18:56:00 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:43.537 18:56:00 -- common/autotest_common.sh@10 -- # set +x 00:05:43.537 ************************************ 00:05:43.537 START TEST env 00:05:43.537 ************************************ 00:05:43.537 18:56:00 env -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:05:43.537 * Looking for test storage... 00:05:43.537 * Found test storage at /home/vagrant/spdk_repo/spdk/test/env 00:05:43.537 18:56:01 env -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:05:43.537 18:56:01 env -- common/autotest_common.sh@1711 -- # lcov --version 00:05:43.537 18:56:01 env -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:05:43.797 18:56:01 env -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:05:43.797 18:56:01 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:43.797 18:56:01 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:43.797 18:56:01 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:43.797 18:56:01 env -- scripts/common.sh@336 -- # IFS=.-: 00:05:43.797 18:56:01 env -- scripts/common.sh@336 -- # read -ra ver1 00:05:43.797 18:56:01 env -- scripts/common.sh@337 -- # IFS=.-: 00:05:43.797 18:56:01 env -- scripts/common.sh@337 -- # read -ra ver2 00:05:43.797 18:56:01 env -- scripts/common.sh@338 -- # local 'op=<' 00:05:43.797 18:56:01 env -- scripts/common.sh@340 -- # ver1_l=2 00:05:43.797 18:56:01 env -- scripts/common.sh@341 -- # ver2_l=1 00:05:43.797 18:56:01 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:43.797 18:56:01 env -- scripts/common.sh@344 -- # case "$op" in 00:05:43.797 18:56:01 env -- scripts/common.sh@345 -- # : 1 00:05:43.797 18:56:01 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:43.797 18:56:01 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:43.797 18:56:01 env -- scripts/common.sh@365 -- # decimal 1 00:05:43.797 18:56:01 env -- scripts/common.sh@353 -- # local d=1 00:05:43.797 18:56:01 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:43.797 18:56:01 env -- scripts/common.sh@355 -- # echo 1 00:05:43.797 18:56:01 env -- scripts/common.sh@365 -- # ver1[v]=1 00:05:43.797 18:56:01 env -- scripts/common.sh@366 -- # decimal 2 00:05:43.797 18:56:01 env -- scripts/common.sh@353 -- # local d=2 00:05:43.797 18:56:01 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:43.797 18:56:01 env -- scripts/common.sh@355 -- # echo 2 00:05:43.797 18:56:01 env -- scripts/common.sh@366 -- # ver2[v]=2 00:05:43.797 18:56:01 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:43.797 18:56:01 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:43.797 18:56:01 env -- scripts/common.sh@368 -- # return 0 00:05:43.797 18:56:01 env -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:43.797 18:56:01 env -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:05:43.797 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:43.797 --rc genhtml_branch_coverage=1 00:05:43.797 --rc genhtml_function_coverage=1 00:05:43.797 --rc genhtml_legend=1 00:05:43.797 --rc geninfo_all_blocks=1 00:05:43.797 --rc geninfo_unexecuted_blocks=1 00:05:43.797 00:05:43.797 ' 00:05:43.797 18:56:01 env -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:05:43.797 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:43.797 --rc genhtml_branch_coverage=1 00:05:43.797 --rc genhtml_function_coverage=1 00:05:43.797 --rc genhtml_legend=1 00:05:43.797 --rc geninfo_all_blocks=1 00:05:43.797 --rc geninfo_unexecuted_blocks=1 00:05:43.797 00:05:43.797 ' 00:05:43.797 18:56:01 env -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:05:43.797 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:43.797 --rc genhtml_branch_coverage=1 00:05:43.797 --rc genhtml_function_coverage=1 00:05:43.797 --rc genhtml_legend=1 00:05:43.797 --rc geninfo_all_blocks=1 00:05:43.797 --rc geninfo_unexecuted_blocks=1 00:05:43.797 00:05:43.797 ' 00:05:43.797 18:56:01 env -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:05:43.797 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:43.797 --rc genhtml_branch_coverage=1 00:05:43.797 --rc genhtml_function_coverage=1 00:05:43.797 --rc genhtml_legend=1 00:05:43.797 --rc geninfo_all_blocks=1 00:05:43.797 --rc geninfo_unexecuted_blocks=1 00:05:43.797 00:05:43.797 ' 00:05:43.797 18:56:01 env -- env/env.sh@10 -- # run_test env_memory /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:05:43.797 18:56:01 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:43.797 18:56:01 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:43.797 18:56:01 env -- common/autotest_common.sh@10 -- # set +x 00:05:43.797 ************************************ 00:05:43.797 START TEST env_memory 00:05:43.797 ************************************ 00:05:43.797 18:56:01 env.env_memory -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:05:43.797 00:05:43.797 00:05:43.797 CUnit - A unit testing framework for C - Version 2.1-3 00:05:43.797 http://cunit.sourceforge.net/ 00:05:43.797 00:05:43.797 00:05:43.797 Suite: memory 00:05:43.797 Test: alloc and free memory map ...[2024-12-05 18:56:01.199003] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:05:43.797 passed 00:05:43.797 Test: mem map translation ...[2024-12-05 18:56:01.238762] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:05:43.797 [2024-12-05 18:56:01.238798] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:05:43.797 [2024-12-05 18:56:01.238848] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:05:43.797 [2024-12-05 18:56:01.238868] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:05:43.797 passed 00:05:43.797 Test: mem map registration ...[2024-12-05 18:56:01.299788] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:05:43.797 [2024-12-05 18:56:01.299823] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:05:43.797 passed 00:05:44.057 Test: mem map adjacent registrations ...passed 00:05:44.057 00:05:44.057 Run Summary: Type Total Ran Passed Failed Inactive 00:05:44.057 suites 1 1 n/a 0 0 00:05:44.057 tests 4 4 4 0 0 00:05:44.057 asserts 152 152 152 0 n/a 00:05:44.057 00:05:44.057 Elapsed time = 0.221 seconds 00:05:44.057 00:05:44.057 real 0m0.270s 00:05:44.057 user 0m0.233s 00:05:44.057 sys 0m0.027s 00:05:44.057 18:56:01 env.env_memory -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:44.057 18:56:01 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:05:44.057 ************************************ 00:05:44.057 END TEST env_memory 00:05:44.057 ************************************ 00:05:44.057 18:56:01 env -- env/env.sh@11 -- # run_test env_vtophys /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:05:44.057 18:56:01 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:44.057 18:56:01 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:44.057 18:56:01 env -- common/autotest_common.sh@10 -- # set +x 00:05:44.057 ************************************ 00:05:44.057 START TEST env_vtophys 00:05:44.057 ************************************ 00:05:44.057 18:56:01 env.env_vtophys -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:05:44.057 EAL: lib.eal log level changed from notice to debug 00:05:44.057 EAL: Detected lcore 0 as core 0 on socket 0 00:05:44.057 EAL: Detected lcore 1 as core 0 on socket 0 00:05:44.057 EAL: Detected lcore 2 as core 0 on socket 0 00:05:44.057 EAL: Detected lcore 3 as core 0 on socket 0 00:05:44.057 EAL: Detected lcore 4 as core 0 on socket 0 00:05:44.057 EAL: Detected lcore 5 as core 0 on socket 0 00:05:44.057 EAL: Detected lcore 6 as core 0 on socket 0 00:05:44.057 EAL: Detected lcore 7 as core 0 on socket 0 00:05:44.057 EAL: Detected lcore 8 as core 0 on socket 0 00:05:44.057 EAL: Detected lcore 9 as core 0 on socket 0 00:05:44.057 EAL: Maximum logical cores by configuration: 128 00:05:44.057 EAL: Detected CPU lcores: 10 00:05:44.057 EAL: Detected NUMA nodes: 1 00:05:44.057 EAL: Checking presence of .so 'librte_eal.so.23.0' 00:05:44.057 EAL: Detected shared linkage of DPDK 00:05:44.057 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so.23.0 00:05:44.057 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so.23.0 00:05:44.057 EAL: Registered [vdev] bus. 00:05:44.057 EAL: bus.vdev log level changed from disabled to notice 00:05:44.057 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so.23.0 00:05:44.057 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so.23.0 00:05:44.057 EAL: pmd.net.i40e.init log level changed from disabled to notice 00:05:44.057 EAL: pmd.net.i40e.driver log level changed from disabled to notice 00:05:44.057 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so 00:05:44.057 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so 00:05:44.057 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so 00:05:44.057 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so 00:05:44.057 EAL: No shared files mode enabled, IPC will be disabled 00:05:44.057 EAL: No shared files mode enabled, IPC is disabled 00:05:44.057 EAL: Selected IOVA mode 'PA' 00:05:44.057 EAL: Probing VFIO support... 00:05:44.057 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:05:44.057 EAL: VFIO modules not loaded, skipping VFIO support... 00:05:44.057 EAL: Ask a virtual area of 0x2e000 bytes 00:05:44.057 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:05:44.057 EAL: Setting up physically contiguous memory... 00:05:44.057 EAL: Setting maximum number of open files to 524288 00:05:44.057 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:05:44.057 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:05:44.057 EAL: Ask a virtual area of 0x61000 bytes 00:05:44.057 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:05:44.057 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:44.057 EAL: Ask a virtual area of 0x400000000 bytes 00:05:44.057 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:05:44.057 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:05:44.057 EAL: Ask a virtual area of 0x61000 bytes 00:05:44.057 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:05:44.057 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:44.057 EAL: Ask a virtual area of 0x400000000 bytes 00:05:44.057 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:05:44.057 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:05:44.057 EAL: Ask a virtual area of 0x61000 bytes 00:05:44.057 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:05:44.057 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:44.057 EAL: Ask a virtual area of 0x400000000 bytes 00:05:44.057 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:05:44.057 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:05:44.057 EAL: Ask a virtual area of 0x61000 bytes 00:05:44.057 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:05:44.057 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:44.057 EAL: Ask a virtual area of 0x400000000 bytes 00:05:44.057 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:05:44.057 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:05:44.057 EAL: Hugepages will be freed exactly as allocated. 00:05:44.057 EAL: No shared files mode enabled, IPC is disabled 00:05:44.057 EAL: No shared files mode enabled, IPC is disabled 00:05:44.316 EAL: TSC frequency is ~2290000 KHz 00:05:44.316 EAL: Main lcore 0 is ready (tid=7fb754e37a40;cpuset=[0]) 00:05:44.316 EAL: Trying to obtain current memory policy. 00:05:44.316 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:44.316 EAL: Restoring previous memory policy: 0 00:05:44.316 EAL: request: mp_malloc_sync 00:05:44.316 EAL: No shared files mode enabled, IPC is disabled 00:05:44.316 EAL: Heap on socket 0 was expanded by 2MB 00:05:44.316 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:05:44.316 EAL: No shared files mode enabled, IPC is disabled 00:05:44.316 EAL: No PCI address specified using 'addr=' in: bus=pci 00:05:44.316 EAL: Mem event callback 'spdk:(nil)' registered 00:05:44.316 EAL: Module /sys/module/vfio_pci not found! error 2 (No such file or directory) 00:05:44.316 00:05:44.316 00:05:44.316 CUnit - A unit testing framework for C - Version 2.1-3 00:05:44.316 http://cunit.sourceforge.net/ 00:05:44.316 00:05:44.316 00:05:44.316 Suite: components_suite 00:05:44.574 Test: vtophys_malloc_test ...passed 00:05:44.574 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:05:44.574 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:44.574 EAL: Restoring previous memory policy: 4 00:05:44.574 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.574 EAL: request: mp_malloc_sync 00:05:44.574 EAL: No shared files mode enabled, IPC is disabled 00:05:44.574 EAL: Heap on socket 0 was expanded by 4MB 00:05:44.574 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.574 EAL: request: mp_malloc_sync 00:05:44.574 EAL: No shared files mode enabled, IPC is disabled 00:05:44.574 EAL: Heap on socket 0 was shrunk by 4MB 00:05:44.574 EAL: Trying to obtain current memory policy. 00:05:44.574 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:44.574 EAL: Restoring previous memory policy: 4 00:05:44.574 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.574 EAL: request: mp_malloc_sync 00:05:44.574 EAL: No shared files mode enabled, IPC is disabled 00:05:44.574 EAL: Heap on socket 0 was expanded by 6MB 00:05:44.574 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.574 EAL: request: mp_malloc_sync 00:05:44.574 EAL: No shared files mode enabled, IPC is disabled 00:05:44.574 EAL: Heap on socket 0 was shrunk by 6MB 00:05:44.574 EAL: Trying to obtain current memory policy. 00:05:44.574 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:44.574 EAL: Restoring previous memory policy: 4 00:05:44.574 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.574 EAL: request: mp_malloc_sync 00:05:44.574 EAL: No shared files mode enabled, IPC is disabled 00:05:44.574 EAL: Heap on socket 0 was expanded by 10MB 00:05:44.574 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.574 EAL: request: mp_malloc_sync 00:05:44.574 EAL: No shared files mode enabled, IPC is disabled 00:05:44.574 EAL: Heap on socket 0 was shrunk by 10MB 00:05:44.574 EAL: Trying to obtain current memory policy. 00:05:44.574 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:44.574 EAL: Restoring previous memory policy: 4 00:05:44.574 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.574 EAL: request: mp_malloc_sync 00:05:44.574 EAL: No shared files mode enabled, IPC is disabled 00:05:44.574 EAL: Heap on socket 0 was expanded by 18MB 00:05:44.574 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.574 EAL: request: mp_malloc_sync 00:05:44.574 EAL: No shared files mode enabled, IPC is disabled 00:05:44.574 EAL: Heap on socket 0 was shrunk by 18MB 00:05:44.574 EAL: Trying to obtain current memory policy. 00:05:44.574 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:44.574 EAL: Restoring previous memory policy: 4 00:05:44.574 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.574 EAL: request: mp_malloc_sync 00:05:44.574 EAL: No shared files mode enabled, IPC is disabled 00:05:44.574 EAL: Heap on socket 0 was expanded by 34MB 00:05:44.574 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.574 EAL: request: mp_malloc_sync 00:05:44.575 EAL: No shared files mode enabled, IPC is disabled 00:05:44.575 EAL: Heap on socket 0 was shrunk by 34MB 00:05:44.575 EAL: Trying to obtain current memory policy. 00:05:44.575 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:44.575 EAL: Restoring previous memory policy: 4 00:05:44.575 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.575 EAL: request: mp_malloc_sync 00:05:44.575 EAL: No shared files mode enabled, IPC is disabled 00:05:44.575 EAL: Heap on socket 0 was expanded by 66MB 00:05:44.575 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.575 EAL: request: mp_malloc_sync 00:05:44.575 EAL: No shared files mode enabled, IPC is disabled 00:05:44.575 EAL: Heap on socket 0 was shrunk by 66MB 00:05:44.575 EAL: Trying to obtain current memory policy. 00:05:44.575 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:44.575 EAL: Restoring previous memory policy: 4 00:05:44.575 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.575 EAL: request: mp_malloc_sync 00:05:44.575 EAL: No shared files mode enabled, IPC is disabled 00:05:44.575 EAL: Heap on socket 0 was expanded by 130MB 00:05:44.575 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.575 EAL: request: mp_malloc_sync 00:05:44.575 EAL: No shared files mode enabled, IPC is disabled 00:05:44.575 EAL: Heap on socket 0 was shrunk by 130MB 00:05:44.575 EAL: Trying to obtain current memory policy. 00:05:44.575 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:44.834 EAL: Restoring previous memory policy: 4 00:05:44.834 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.834 EAL: request: mp_malloc_sync 00:05:44.834 EAL: No shared files mode enabled, IPC is disabled 00:05:44.834 EAL: Heap on socket 0 was expanded by 258MB 00:05:44.834 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.834 EAL: request: mp_malloc_sync 00:05:44.834 EAL: No shared files mode enabled, IPC is disabled 00:05:44.834 EAL: Heap on socket 0 was shrunk by 258MB 00:05:44.834 EAL: Trying to obtain current memory policy. 00:05:44.834 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:44.834 EAL: Restoring previous memory policy: 4 00:05:44.834 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.834 EAL: request: mp_malloc_sync 00:05:44.834 EAL: No shared files mode enabled, IPC is disabled 00:05:44.834 EAL: Heap on socket 0 was expanded by 514MB 00:05:45.094 EAL: Calling mem event callback 'spdk:(nil)' 00:05:45.094 EAL: request: mp_malloc_sync 00:05:45.094 EAL: No shared files mode enabled, IPC is disabled 00:05:45.094 EAL: Heap on socket 0 was shrunk by 514MB 00:05:45.094 EAL: Trying to obtain current memory policy. 00:05:45.094 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:45.354 EAL: Restoring previous memory policy: 4 00:05:45.354 EAL: Calling mem event callback 'spdk:(nil)' 00:05:45.354 EAL: request: mp_malloc_sync 00:05:45.354 EAL: No shared files mode enabled, IPC is disabled 00:05:45.354 EAL: Heap on socket 0 was expanded by 1026MB 00:05:45.615 EAL: Calling mem event callback 'spdk:(nil)' 00:05:45.615 passed 00:05:45.615 00:05:45.615 Run Summary: Type Total Ran Passed Failed Inactive 00:05:45.615 suites 1 1 n/a 0 0 00:05:45.615 tests 2 2 2 0 0 00:05:45.615 asserts 5337 5337 5337 0 n/a 00:05:45.615 00:05:45.615 Elapsed time = 1.392 seconds 00:05:45.615 EAL: request: mp_malloc_sync 00:05:45.615 EAL: No shared files mode enabled, IPC is disabled 00:05:45.615 EAL: Heap on socket 0 was shrunk by 1026MB 00:05:45.615 EAL: Calling mem event callback 'spdk:(nil)' 00:05:45.615 EAL: request: mp_malloc_sync 00:05:45.615 EAL: No shared files mode enabled, IPC is disabled 00:05:45.615 EAL: Heap on socket 0 was shrunk by 2MB 00:05:45.615 EAL: No shared files mode enabled, IPC is disabled 00:05:45.615 EAL: No shared files mode enabled, IPC is disabled 00:05:45.615 EAL: No shared files mode enabled, IPC is disabled 00:05:45.615 ************************************ 00:05:45.615 END TEST env_vtophys 00:05:45.615 00:05:45.615 real 0m1.647s 00:05:45.615 user 0m0.804s 00:05:45.615 sys 0m0.706s 00:05:45.615 18:56:03 env.env_vtophys -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:45.615 18:56:03 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:05:45.615 ************************************ 00:05:45.875 18:56:03 env -- env/env.sh@12 -- # run_test env_pci /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:05:45.875 18:56:03 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:45.875 18:56:03 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:45.875 18:56:03 env -- common/autotest_common.sh@10 -- # set +x 00:05:45.875 ************************************ 00:05:45.875 START TEST env_pci 00:05:45.875 ************************************ 00:05:45.875 18:56:03 env.env_pci -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:05:45.875 00:05:45.875 00:05:45.875 CUnit - A unit testing framework for C - Version 2.1-3 00:05:45.875 http://cunit.sourceforge.net/ 00:05:45.875 00:05:45.875 00:05:45.875 Suite: pci 00:05:45.875 Test: pci_hook ...[2024-12-05 18:56:03.223039] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/pci.c:1117:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 68709 has claimed it 00:05:45.875 passed 00:05:45.875 00:05:45.875 Run Summary: Type Total Ran Passed Failed Inactive 00:05:45.875 suites 1 1 n/a 0 0 00:05:45.875 tests 1 1 1 0 0 00:05:45.875 asserts 25 25 25 0 n/a 00:05:45.875 00:05:45.875 Elapsed time = 0.007 seconds 00:05:45.875 EAL: Cannot find device (10000:00:01.0) 00:05:45.875 EAL: Failed to attach device on primary process 00:05:45.875 00:05:45.875 real 0m0.087s 00:05:45.875 user 0m0.039s 00:05:45.875 sys 0m0.047s 00:05:45.875 18:56:03 env.env_pci -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:45.875 18:56:03 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:05:45.875 ************************************ 00:05:45.875 END TEST env_pci 00:05:45.875 ************************************ 00:05:45.875 18:56:03 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:05:45.875 18:56:03 env -- env/env.sh@15 -- # uname 00:05:45.875 18:56:03 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:05:45.875 18:56:03 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:05:45.875 18:56:03 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:05:45.875 18:56:03 env -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:05:45.875 18:56:03 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:45.875 18:56:03 env -- common/autotest_common.sh@10 -- # set +x 00:05:45.875 ************************************ 00:05:45.875 START TEST env_dpdk_post_init 00:05:45.875 ************************************ 00:05:45.875 18:56:03 env.env_dpdk_post_init -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:05:45.875 EAL: Detected CPU lcores: 10 00:05:45.875 EAL: Detected NUMA nodes: 1 00:05:45.875 EAL: Detected shared linkage of DPDK 00:05:45.875 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:05:45.875 EAL: Selected IOVA mode 'PA' 00:05:46.135 TELEMETRY: No legacy callbacks, legacy socket not created 00:05:46.135 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:10.0 (socket -1) 00:05:46.135 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:11.0 (socket -1) 00:05:46.135 Starting DPDK initialization... 00:05:46.135 Starting SPDK post initialization... 00:05:46.135 SPDK NVMe probe 00:05:46.135 Attaching to 0000:00:10.0 00:05:46.135 Attaching to 0000:00:11.0 00:05:46.135 Attached to 0000:00:10.0 00:05:46.135 Attached to 0000:00:11.0 00:05:46.135 Cleaning up... 00:05:46.135 00:05:46.135 real 0m0.232s 00:05:46.135 user 0m0.063s 00:05:46.135 sys 0m0.071s 00:05:46.135 18:56:03 env.env_dpdk_post_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:46.135 ************************************ 00:05:46.135 END TEST env_dpdk_post_init 00:05:46.135 ************************************ 00:05:46.135 18:56:03 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:05:46.135 18:56:03 env -- env/env.sh@26 -- # uname 00:05:46.135 18:56:03 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:05:46.135 18:56:03 env -- env/env.sh@29 -- # run_test env_mem_callbacks /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:05:46.135 18:56:03 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:46.135 18:56:03 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:46.135 18:56:03 env -- common/autotest_common.sh@10 -- # set +x 00:05:46.135 ************************************ 00:05:46.135 START TEST env_mem_callbacks 00:05:46.135 ************************************ 00:05:46.135 18:56:03 env.env_mem_callbacks -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:05:46.395 EAL: Detected CPU lcores: 10 00:05:46.395 EAL: Detected NUMA nodes: 1 00:05:46.395 EAL: Detected shared linkage of DPDK 00:05:46.395 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:05:46.395 EAL: Selected IOVA mode 'PA' 00:05:46.396 TELEMETRY: No legacy callbacks, legacy socket not created 00:05:46.396 00:05:46.396 00:05:46.396 CUnit - A unit testing framework for C - Version 2.1-3 00:05:46.396 http://cunit.sourceforge.net/ 00:05:46.396 00:05:46.396 00:05:46.396 Suite: memory 00:05:46.396 Test: test ... 00:05:46.396 register 0x200000200000 2097152 00:05:46.396 malloc 3145728 00:05:46.396 register 0x200000400000 4194304 00:05:46.396 buf 0x200000500000 len 3145728 PASSED 00:05:46.396 malloc 64 00:05:46.396 buf 0x2000004fff40 len 64 PASSED 00:05:46.396 malloc 4194304 00:05:46.396 register 0x200000800000 6291456 00:05:46.396 buf 0x200000a00000 len 4194304 PASSED 00:05:46.396 free 0x200000500000 3145728 00:05:46.396 free 0x2000004fff40 64 00:05:46.396 unregister 0x200000400000 4194304 PASSED 00:05:46.396 free 0x200000a00000 4194304 00:05:46.396 unregister 0x200000800000 6291456 PASSED 00:05:46.396 malloc 8388608 00:05:46.396 register 0x200000400000 10485760 00:05:46.396 buf 0x200000600000 len 8388608 PASSED 00:05:46.396 free 0x200000600000 8388608 00:05:46.396 unregister 0x200000400000 10485760 PASSED 00:05:46.396 passed 00:05:46.396 00:05:46.396 Run Summary: Type Total Ran Passed Failed Inactive 00:05:46.396 suites 1 1 n/a 0 0 00:05:46.396 tests 1 1 1 0 0 00:05:46.396 asserts 15 15 15 0 n/a 00:05:46.396 00:05:46.396 Elapsed time = 0.011 seconds 00:05:46.396 00:05:46.396 real 0m0.182s 00:05:46.396 user 0m0.021s 00:05:46.396 sys 0m0.057s 00:05:46.396 18:56:03 env.env_mem_callbacks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:46.396 18:56:03 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:05:46.396 ************************************ 00:05:46.396 END TEST env_mem_callbacks 00:05:46.396 ************************************ 00:05:46.396 00:05:46.396 real 0m3.003s 00:05:46.396 user 0m1.383s 00:05:46.396 sys 0m1.266s 00:05:46.396 18:56:03 env -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:46.396 18:56:03 env -- common/autotest_common.sh@10 -- # set +x 00:05:46.396 ************************************ 00:05:46.396 END TEST env 00:05:46.396 ************************************ 00:05:46.656 18:56:03 -- spdk/autotest.sh@156 -- # run_test rpc /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:05:46.656 18:56:03 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:46.656 18:56:03 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:46.656 18:56:03 -- common/autotest_common.sh@10 -- # set +x 00:05:46.656 ************************************ 00:05:46.656 START TEST rpc 00:05:46.656 ************************************ 00:05:46.656 18:56:03 rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:05:46.656 * Looking for test storage... 00:05:46.656 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:05:46.656 18:56:04 rpc -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:05:46.656 18:56:04 rpc -- common/autotest_common.sh@1711 -- # lcov --version 00:05:46.656 18:56:04 rpc -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:05:46.656 18:56:04 rpc -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:05:46.656 18:56:04 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:46.656 18:56:04 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:46.656 18:56:04 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:46.656 18:56:04 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:46.656 18:56:04 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:46.656 18:56:04 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:46.656 18:56:04 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:46.656 18:56:04 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:46.656 18:56:04 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:46.656 18:56:04 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:46.656 18:56:04 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:46.656 18:56:04 rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:46.656 18:56:04 rpc -- scripts/common.sh@345 -- # : 1 00:05:46.656 18:56:04 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:46.656 18:56:04 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:46.656 18:56:04 rpc -- scripts/common.sh@365 -- # decimal 1 00:05:46.656 18:56:04 rpc -- scripts/common.sh@353 -- # local d=1 00:05:46.656 18:56:04 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:46.656 18:56:04 rpc -- scripts/common.sh@355 -- # echo 1 00:05:46.656 18:56:04 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:46.656 18:56:04 rpc -- scripts/common.sh@366 -- # decimal 2 00:05:46.656 18:56:04 rpc -- scripts/common.sh@353 -- # local d=2 00:05:46.656 18:56:04 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:46.656 18:56:04 rpc -- scripts/common.sh@355 -- # echo 2 00:05:46.656 18:56:04 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:46.656 18:56:04 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:46.656 18:56:04 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:46.656 18:56:04 rpc -- scripts/common.sh@368 -- # return 0 00:05:46.656 18:56:04 rpc -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:46.656 18:56:04 rpc -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:05:46.656 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.656 --rc genhtml_branch_coverage=1 00:05:46.656 --rc genhtml_function_coverage=1 00:05:46.656 --rc genhtml_legend=1 00:05:46.656 --rc geninfo_all_blocks=1 00:05:46.656 --rc geninfo_unexecuted_blocks=1 00:05:46.656 00:05:46.656 ' 00:05:46.656 18:56:04 rpc -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:05:46.656 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.656 --rc genhtml_branch_coverage=1 00:05:46.656 --rc genhtml_function_coverage=1 00:05:46.656 --rc genhtml_legend=1 00:05:46.656 --rc geninfo_all_blocks=1 00:05:46.656 --rc geninfo_unexecuted_blocks=1 00:05:46.656 00:05:46.656 ' 00:05:46.656 18:56:04 rpc -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:05:46.656 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.656 --rc genhtml_branch_coverage=1 00:05:46.656 --rc genhtml_function_coverage=1 00:05:46.656 --rc genhtml_legend=1 00:05:46.656 --rc geninfo_all_blocks=1 00:05:46.656 --rc geninfo_unexecuted_blocks=1 00:05:46.656 00:05:46.656 ' 00:05:46.656 18:56:04 rpc -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:05:46.656 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.656 --rc genhtml_branch_coverage=1 00:05:46.656 --rc genhtml_function_coverage=1 00:05:46.656 --rc genhtml_legend=1 00:05:46.656 --rc geninfo_all_blocks=1 00:05:46.656 --rc geninfo_unexecuted_blocks=1 00:05:46.656 00:05:46.656 ' 00:05:46.656 18:56:04 rpc -- rpc/rpc.sh@65 -- # spdk_pid=68836 00:05:46.656 18:56:04 rpc -- rpc/rpc.sh@64 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -e bdev 00:05:46.656 18:56:04 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:46.656 18:56:04 rpc -- rpc/rpc.sh@67 -- # waitforlisten 68836 00:05:46.656 18:56:04 rpc -- common/autotest_common.sh@835 -- # '[' -z 68836 ']' 00:05:46.656 18:56:04 rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:46.656 18:56:04 rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:46.656 18:56:04 rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:46.656 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:46.656 18:56:04 rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:46.656 18:56:04 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:46.916 [2024-12-05 18:56:04.294147] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:05:46.916 [2024-12-05 18:56:04.294388] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid68836 ] 00:05:46.916 [2024-12-05 18:56:04.451850] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:47.175 [2024-12-05 18:56:04.477420] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:05:47.175 [2024-12-05 18:56:04.477474] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 68836' to capture a snapshot of events at runtime. 00:05:47.175 [2024-12-05 18:56:04.477486] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:05:47.175 [2024-12-05 18:56:04.477502] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:05:47.175 [2024-12-05 18:56:04.477512] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid68836 for offline analysis/debug. 00:05:47.175 [2024-12-05 18:56:04.477901] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:47.744 18:56:05 rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:47.744 18:56:05 rpc -- common/autotest_common.sh@868 -- # return 0 00:05:47.744 18:56:05 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:05:47.744 18:56:05 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:05:47.744 18:56:05 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:05:47.744 18:56:05 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:05:47.744 18:56:05 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:47.744 18:56:05 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:47.744 18:56:05 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:47.744 ************************************ 00:05:47.744 START TEST rpc_integrity 00:05:47.744 ************************************ 00:05:47.744 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:05:47.744 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:05:47.744 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:47.744 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.744 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:47.744 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:05:47.744 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:05:47.744 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:05:47.744 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:05:47.744 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:47.744 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.744 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:47.744 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:05:47.744 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:05:47.744 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:47.744 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.744 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:47.744 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:05:47.744 { 00:05:47.744 "name": "Malloc0", 00:05:47.744 "aliases": [ 00:05:47.744 "eace9352-b5dd-4d31-b731-51f2cc4e22d9" 00:05:47.744 ], 00:05:47.744 "product_name": "Malloc disk", 00:05:47.744 "block_size": 512, 00:05:47.744 "num_blocks": 16384, 00:05:47.744 "uuid": "eace9352-b5dd-4d31-b731-51f2cc4e22d9", 00:05:47.744 "assigned_rate_limits": { 00:05:47.744 "rw_ios_per_sec": 0, 00:05:47.744 "rw_mbytes_per_sec": 0, 00:05:47.744 "r_mbytes_per_sec": 0, 00:05:47.744 "w_mbytes_per_sec": 0 00:05:47.744 }, 00:05:47.744 "claimed": false, 00:05:47.744 "zoned": false, 00:05:47.744 "supported_io_types": { 00:05:47.744 "read": true, 00:05:47.744 "write": true, 00:05:47.744 "unmap": true, 00:05:47.744 "flush": true, 00:05:47.744 "reset": true, 00:05:47.744 "nvme_admin": false, 00:05:47.744 "nvme_io": false, 00:05:47.744 "nvme_io_md": false, 00:05:47.744 "write_zeroes": true, 00:05:47.744 "zcopy": true, 00:05:47.744 "get_zone_info": false, 00:05:47.744 "zone_management": false, 00:05:47.744 "zone_append": false, 00:05:47.744 "compare": false, 00:05:47.744 "compare_and_write": false, 00:05:47.744 "abort": true, 00:05:47.744 "seek_hole": false, 00:05:47.744 "seek_data": false, 00:05:47.744 "copy": true, 00:05:47.744 "nvme_iov_md": false 00:05:47.744 }, 00:05:47.744 "memory_domains": [ 00:05:47.744 { 00:05:47.744 "dma_device_id": "system", 00:05:47.744 "dma_device_type": 1 00:05:47.744 }, 00:05:47.744 { 00:05:47.744 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:47.744 "dma_device_type": 2 00:05:47.744 } 00:05:47.744 ], 00:05:47.744 "driver_specific": {} 00:05:47.744 } 00:05:47.744 ]' 00:05:47.744 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:05:47.744 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:05:47.744 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:05:47.744 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:47.744 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.744 [2024-12-05 18:56:05.248247] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:05:47.744 [2024-12-05 18:56:05.248307] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:05:47.744 [2024-12-05 18:56:05.248335] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006c80 00:05:47.744 [2024-12-05 18:56:05.248344] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:05:47.745 [2024-12-05 18:56:05.250571] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:05:47.745 [2024-12-05 18:56:05.250604] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:05:47.745 Passthru0 00:05:47.745 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:47.745 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:05:47.745 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:47.745 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.745 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:47.745 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:05:47.745 { 00:05:47.745 "name": "Malloc0", 00:05:47.745 "aliases": [ 00:05:47.745 "eace9352-b5dd-4d31-b731-51f2cc4e22d9" 00:05:47.745 ], 00:05:47.745 "product_name": "Malloc disk", 00:05:47.745 "block_size": 512, 00:05:47.745 "num_blocks": 16384, 00:05:47.745 "uuid": "eace9352-b5dd-4d31-b731-51f2cc4e22d9", 00:05:47.745 "assigned_rate_limits": { 00:05:47.745 "rw_ios_per_sec": 0, 00:05:47.745 "rw_mbytes_per_sec": 0, 00:05:47.745 "r_mbytes_per_sec": 0, 00:05:47.745 "w_mbytes_per_sec": 0 00:05:47.745 }, 00:05:47.745 "claimed": true, 00:05:47.745 "claim_type": "exclusive_write", 00:05:47.745 "zoned": false, 00:05:47.745 "supported_io_types": { 00:05:47.745 "read": true, 00:05:47.745 "write": true, 00:05:47.745 "unmap": true, 00:05:47.745 "flush": true, 00:05:47.745 "reset": true, 00:05:47.745 "nvme_admin": false, 00:05:47.745 "nvme_io": false, 00:05:47.745 "nvme_io_md": false, 00:05:47.745 "write_zeroes": true, 00:05:47.745 "zcopy": true, 00:05:47.745 "get_zone_info": false, 00:05:47.745 "zone_management": false, 00:05:47.745 "zone_append": false, 00:05:47.745 "compare": false, 00:05:47.745 "compare_and_write": false, 00:05:47.745 "abort": true, 00:05:47.745 "seek_hole": false, 00:05:47.745 "seek_data": false, 00:05:47.745 "copy": true, 00:05:47.745 "nvme_iov_md": false 00:05:47.745 }, 00:05:47.745 "memory_domains": [ 00:05:47.745 { 00:05:47.745 "dma_device_id": "system", 00:05:47.745 "dma_device_type": 1 00:05:47.745 }, 00:05:47.745 { 00:05:47.745 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:47.745 "dma_device_type": 2 00:05:47.745 } 00:05:47.745 ], 00:05:47.745 "driver_specific": {} 00:05:47.745 }, 00:05:47.745 { 00:05:47.745 "name": "Passthru0", 00:05:47.745 "aliases": [ 00:05:47.745 "4a195943-15fd-581a-8733-b2fbe9daa2cf" 00:05:47.745 ], 00:05:47.745 "product_name": "passthru", 00:05:47.745 "block_size": 512, 00:05:47.745 "num_blocks": 16384, 00:05:47.745 "uuid": "4a195943-15fd-581a-8733-b2fbe9daa2cf", 00:05:47.745 "assigned_rate_limits": { 00:05:47.745 "rw_ios_per_sec": 0, 00:05:47.745 "rw_mbytes_per_sec": 0, 00:05:47.745 "r_mbytes_per_sec": 0, 00:05:47.745 "w_mbytes_per_sec": 0 00:05:47.745 }, 00:05:47.745 "claimed": false, 00:05:47.745 "zoned": false, 00:05:47.745 "supported_io_types": { 00:05:47.745 "read": true, 00:05:47.745 "write": true, 00:05:47.745 "unmap": true, 00:05:47.745 "flush": true, 00:05:47.745 "reset": true, 00:05:47.745 "nvme_admin": false, 00:05:47.745 "nvme_io": false, 00:05:47.745 "nvme_io_md": false, 00:05:47.745 "write_zeroes": true, 00:05:47.745 "zcopy": true, 00:05:47.745 "get_zone_info": false, 00:05:47.745 "zone_management": false, 00:05:47.745 "zone_append": false, 00:05:47.745 "compare": false, 00:05:47.745 "compare_and_write": false, 00:05:47.745 "abort": true, 00:05:47.745 "seek_hole": false, 00:05:47.745 "seek_data": false, 00:05:47.745 "copy": true, 00:05:47.745 "nvme_iov_md": false 00:05:47.745 }, 00:05:47.745 "memory_domains": [ 00:05:47.745 { 00:05:47.745 "dma_device_id": "system", 00:05:47.745 "dma_device_type": 1 00:05:47.745 }, 00:05:47.745 { 00:05:47.745 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:47.745 "dma_device_type": 2 00:05:47.745 } 00:05:47.745 ], 00:05:47.745 "driver_specific": { 00:05:47.745 "passthru": { 00:05:47.745 "name": "Passthru0", 00:05:47.745 "base_bdev_name": "Malloc0" 00:05:47.745 } 00:05:47.745 } 00:05:47.745 } 00:05:47.745 ]' 00:05:47.745 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:05:48.004 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:05:48.005 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:05:48.005 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.005 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.005 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.005 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:05:48.005 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.005 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.005 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.005 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:05:48.005 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.005 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.005 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.005 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:05:48.005 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:05:48.005 18:56:05 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:05:48.005 00:05:48.005 real 0m0.308s 00:05:48.005 user 0m0.191s 00:05:48.005 sys 0m0.042s 00:05:48.005 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:48.005 18:56:05 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.005 ************************************ 00:05:48.005 END TEST rpc_integrity 00:05:48.005 ************************************ 00:05:48.005 18:56:05 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:05:48.005 18:56:05 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:48.005 18:56:05 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:48.005 18:56:05 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:48.005 ************************************ 00:05:48.005 START TEST rpc_plugins 00:05:48.005 ************************************ 00:05:48.005 18:56:05 rpc.rpc_plugins -- common/autotest_common.sh@1129 -- # rpc_plugins 00:05:48.005 18:56:05 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:05:48.005 18:56:05 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.005 18:56:05 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:48.005 18:56:05 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.005 18:56:05 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:05:48.005 18:56:05 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:05:48.005 18:56:05 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.005 18:56:05 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:48.005 18:56:05 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.005 18:56:05 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:05:48.005 { 00:05:48.005 "name": "Malloc1", 00:05:48.005 "aliases": [ 00:05:48.005 "8b2e0a61-51d4-4baa-81df-a1f023b2d79f" 00:05:48.005 ], 00:05:48.005 "product_name": "Malloc disk", 00:05:48.005 "block_size": 4096, 00:05:48.005 "num_blocks": 256, 00:05:48.005 "uuid": "8b2e0a61-51d4-4baa-81df-a1f023b2d79f", 00:05:48.005 "assigned_rate_limits": { 00:05:48.005 "rw_ios_per_sec": 0, 00:05:48.005 "rw_mbytes_per_sec": 0, 00:05:48.005 "r_mbytes_per_sec": 0, 00:05:48.005 "w_mbytes_per_sec": 0 00:05:48.005 }, 00:05:48.005 "claimed": false, 00:05:48.005 "zoned": false, 00:05:48.005 "supported_io_types": { 00:05:48.005 "read": true, 00:05:48.005 "write": true, 00:05:48.005 "unmap": true, 00:05:48.005 "flush": true, 00:05:48.005 "reset": true, 00:05:48.005 "nvme_admin": false, 00:05:48.005 "nvme_io": false, 00:05:48.005 "nvme_io_md": false, 00:05:48.005 "write_zeroes": true, 00:05:48.005 "zcopy": true, 00:05:48.005 "get_zone_info": false, 00:05:48.005 "zone_management": false, 00:05:48.005 "zone_append": false, 00:05:48.005 "compare": false, 00:05:48.005 "compare_and_write": false, 00:05:48.005 "abort": true, 00:05:48.005 "seek_hole": false, 00:05:48.005 "seek_data": false, 00:05:48.005 "copy": true, 00:05:48.005 "nvme_iov_md": false 00:05:48.005 }, 00:05:48.005 "memory_domains": [ 00:05:48.005 { 00:05:48.005 "dma_device_id": "system", 00:05:48.005 "dma_device_type": 1 00:05:48.005 }, 00:05:48.005 { 00:05:48.005 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:48.005 "dma_device_type": 2 00:05:48.005 } 00:05:48.005 ], 00:05:48.005 "driver_specific": {} 00:05:48.005 } 00:05:48.005 ]' 00:05:48.005 18:56:05 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:05:48.266 18:56:05 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:05:48.266 18:56:05 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:05:48.266 18:56:05 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.266 18:56:05 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:48.266 18:56:05 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.266 18:56:05 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:05:48.266 18:56:05 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.266 18:56:05 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:48.266 18:56:05 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.266 18:56:05 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:05:48.266 18:56:05 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:05:48.266 18:56:05 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:05:48.266 00:05:48.266 real 0m0.155s 00:05:48.266 user 0m0.097s 00:05:48.266 sys 0m0.020s 00:05:48.266 18:56:05 rpc.rpc_plugins -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:48.266 18:56:05 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:48.266 ************************************ 00:05:48.266 END TEST rpc_plugins 00:05:48.266 ************************************ 00:05:48.266 18:56:05 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:05:48.266 18:56:05 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:48.266 18:56:05 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:48.266 18:56:05 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:48.266 ************************************ 00:05:48.266 START TEST rpc_trace_cmd_test 00:05:48.266 ************************************ 00:05:48.266 18:56:05 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1129 -- # rpc_trace_cmd_test 00:05:48.266 18:56:05 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:05:48.266 18:56:05 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:05:48.266 18:56:05 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.266 18:56:05 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:05:48.266 18:56:05 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.266 18:56:05 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:05:48.266 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid68836", 00:05:48.266 "tpoint_group_mask": "0x8", 00:05:48.266 "iscsi_conn": { 00:05:48.266 "mask": "0x2", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 }, 00:05:48.266 "scsi": { 00:05:48.266 "mask": "0x4", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 }, 00:05:48.266 "bdev": { 00:05:48.266 "mask": "0x8", 00:05:48.266 "tpoint_mask": "0xffffffffffffffff" 00:05:48.266 }, 00:05:48.266 "nvmf_rdma": { 00:05:48.266 "mask": "0x10", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 }, 00:05:48.266 "nvmf_tcp": { 00:05:48.266 "mask": "0x20", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 }, 00:05:48.266 "ftl": { 00:05:48.266 "mask": "0x40", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 }, 00:05:48.266 "blobfs": { 00:05:48.266 "mask": "0x80", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 }, 00:05:48.266 "dsa": { 00:05:48.266 "mask": "0x200", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 }, 00:05:48.266 "thread": { 00:05:48.266 "mask": "0x400", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 }, 00:05:48.266 "nvme_pcie": { 00:05:48.266 "mask": "0x800", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 }, 00:05:48.266 "iaa": { 00:05:48.266 "mask": "0x1000", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 }, 00:05:48.266 "nvme_tcp": { 00:05:48.266 "mask": "0x2000", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 }, 00:05:48.266 "bdev_nvme": { 00:05:48.266 "mask": "0x4000", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 }, 00:05:48.266 "sock": { 00:05:48.266 "mask": "0x8000", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 }, 00:05:48.266 "blob": { 00:05:48.266 "mask": "0x10000", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 }, 00:05:48.266 "bdev_raid": { 00:05:48.266 "mask": "0x20000", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 }, 00:05:48.266 "scheduler": { 00:05:48.266 "mask": "0x40000", 00:05:48.266 "tpoint_mask": "0x0" 00:05:48.266 } 00:05:48.266 }' 00:05:48.266 18:56:05 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:05:48.266 18:56:05 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 19 -gt 2 ']' 00:05:48.266 18:56:05 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:05:48.266 18:56:05 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:05:48.266 18:56:05 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:05:48.526 18:56:05 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:05:48.526 18:56:05 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:05:48.526 18:56:05 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:05:48.526 18:56:05 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:05:48.526 18:56:05 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:05:48.526 00:05:48.526 real 0m0.241s 00:05:48.526 user 0m0.203s 00:05:48.526 sys 0m0.029s 00:05:48.526 18:56:05 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:48.526 18:56:05 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:05:48.526 ************************************ 00:05:48.526 END TEST rpc_trace_cmd_test 00:05:48.526 ************************************ 00:05:48.526 18:56:05 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:05:48.526 18:56:05 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:05:48.526 18:56:05 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:05:48.526 18:56:05 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:48.526 18:56:05 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:48.526 18:56:05 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:48.526 ************************************ 00:05:48.526 START TEST rpc_daemon_integrity 00:05:48.526 ************************************ 00:05:48.526 18:56:05 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:05:48.526 18:56:05 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:05:48.526 18:56:05 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.526 18:56:05 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.526 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.526 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:05:48.526 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:05:48.526 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:05:48.526 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:05:48.526 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.526 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.526 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.526 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:05:48.526 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:05:48.526 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.526 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.786 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.786 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:05:48.786 { 00:05:48.786 "name": "Malloc2", 00:05:48.786 "aliases": [ 00:05:48.786 "e9f7a3ea-9d0c-4b65-b367-2c8aaae4ed2c" 00:05:48.786 ], 00:05:48.786 "product_name": "Malloc disk", 00:05:48.786 "block_size": 512, 00:05:48.786 "num_blocks": 16384, 00:05:48.786 "uuid": "e9f7a3ea-9d0c-4b65-b367-2c8aaae4ed2c", 00:05:48.786 "assigned_rate_limits": { 00:05:48.786 "rw_ios_per_sec": 0, 00:05:48.786 "rw_mbytes_per_sec": 0, 00:05:48.786 "r_mbytes_per_sec": 0, 00:05:48.786 "w_mbytes_per_sec": 0 00:05:48.786 }, 00:05:48.786 "claimed": false, 00:05:48.786 "zoned": false, 00:05:48.786 "supported_io_types": { 00:05:48.786 "read": true, 00:05:48.786 "write": true, 00:05:48.786 "unmap": true, 00:05:48.786 "flush": true, 00:05:48.786 "reset": true, 00:05:48.786 "nvme_admin": false, 00:05:48.786 "nvme_io": false, 00:05:48.786 "nvme_io_md": false, 00:05:48.786 "write_zeroes": true, 00:05:48.786 "zcopy": true, 00:05:48.786 "get_zone_info": false, 00:05:48.786 "zone_management": false, 00:05:48.786 "zone_append": false, 00:05:48.786 "compare": false, 00:05:48.786 "compare_and_write": false, 00:05:48.786 "abort": true, 00:05:48.786 "seek_hole": false, 00:05:48.786 "seek_data": false, 00:05:48.786 "copy": true, 00:05:48.786 "nvme_iov_md": false 00:05:48.786 }, 00:05:48.786 "memory_domains": [ 00:05:48.786 { 00:05:48.786 "dma_device_id": "system", 00:05:48.786 "dma_device_type": 1 00:05:48.786 }, 00:05:48.786 { 00:05:48.786 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:48.786 "dma_device_type": 2 00:05:48.786 } 00:05:48.786 ], 00:05:48.786 "driver_specific": {} 00:05:48.786 } 00:05:48.786 ]' 00:05:48.786 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:05:48.786 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:05:48.786 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:05:48.786 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.786 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.786 [2024-12-05 18:56:06.150937] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:05:48.786 [2024-12-05 18:56:06.151000] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:05:48.786 [2024-12-05 18:56:06.151023] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:05:48.786 [2024-12-05 18:56:06.151032] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:05:48.786 [2024-12-05 18:56:06.153415] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:05:48.786 [2024-12-05 18:56:06.153452] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:05:48.786 Passthru0 00:05:48.786 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.786 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:05:48.786 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:05:48.787 { 00:05:48.787 "name": "Malloc2", 00:05:48.787 "aliases": [ 00:05:48.787 "e9f7a3ea-9d0c-4b65-b367-2c8aaae4ed2c" 00:05:48.787 ], 00:05:48.787 "product_name": "Malloc disk", 00:05:48.787 "block_size": 512, 00:05:48.787 "num_blocks": 16384, 00:05:48.787 "uuid": "e9f7a3ea-9d0c-4b65-b367-2c8aaae4ed2c", 00:05:48.787 "assigned_rate_limits": { 00:05:48.787 "rw_ios_per_sec": 0, 00:05:48.787 "rw_mbytes_per_sec": 0, 00:05:48.787 "r_mbytes_per_sec": 0, 00:05:48.787 "w_mbytes_per_sec": 0 00:05:48.787 }, 00:05:48.787 "claimed": true, 00:05:48.787 "claim_type": "exclusive_write", 00:05:48.787 "zoned": false, 00:05:48.787 "supported_io_types": { 00:05:48.787 "read": true, 00:05:48.787 "write": true, 00:05:48.787 "unmap": true, 00:05:48.787 "flush": true, 00:05:48.787 "reset": true, 00:05:48.787 "nvme_admin": false, 00:05:48.787 "nvme_io": false, 00:05:48.787 "nvme_io_md": false, 00:05:48.787 "write_zeroes": true, 00:05:48.787 "zcopy": true, 00:05:48.787 "get_zone_info": false, 00:05:48.787 "zone_management": false, 00:05:48.787 "zone_append": false, 00:05:48.787 "compare": false, 00:05:48.787 "compare_and_write": false, 00:05:48.787 "abort": true, 00:05:48.787 "seek_hole": false, 00:05:48.787 "seek_data": false, 00:05:48.787 "copy": true, 00:05:48.787 "nvme_iov_md": false 00:05:48.787 }, 00:05:48.787 "memory_domains": [ 00:05:48.787 { 00:05:48.787 "dma_device_id": "system", 00:05:48.787 "dma_device_type": 1 00:05:48.787 }, 00:05:48.787 { 00:05:48.787 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:48.787 "dma_device_type": 2 00:05:48.787 } 00:05:48.787 ], 00:05:48.787 "driver_specific": {} 00:05:48.787 }, 00:05:48.787 { 00:05:48.787 "name": "Passthru0", 00:05:48.787 "aliases": [ 00:05:48.787 "94f77fb3-9fc4-5abf-8bab-a6e120435e4f" 00:05:48.787 ], 00:05:48.787 "product_name": "passthru", 00:05:48.787 "block_size": 512, 00:05:48.787 "num_blocks": 16384, 00:05:48.787 "uuid": "94f77fb3-9fc4-5abf-8bab-a6e120435e4f", 00:05:48.787 "assigned_rate_limits": { 00:05:48.787 "rw_ios_per_sec": 0, 00:05:48.787 "rw_mbytes_per_sec": 0, 00:05:48.787 "r_mbytes_per_sec": 0, 00:05:48.787 "w_mbytes_per_sec": 0 00:05:48.787 }, 00:05:48.787 "claimed": false, 00:05:48.787 "zoned": false, 00:05:48.787 "supported_io_types": { 00:05:48.787 "read": true, 00:05:48.787 "write": true, 00:05:48.787 "unmap": true, 00:05:48.787 "flush": true, 00:05:48.787 "reset": true, 00:05:48.787 "nvme_admin": false, 00:05:48.787 "nvme_io": false, 00:05:48.787 "nvme_io_md": false, 00:05:48.787 "write_zeroes": true, 00:05:48.787 "zcopy": true, 00:05:48.787 "get_zone_info": false, 00:05:48.787 "zone_management": false, 00:05:48.787 "zone_append": false, 00:05:48.787 "compare": false, 00:05:48.787 "compare_and_write": false, 00:05:48.787 "abort": true, 00:05:48.787 "seek_hole": false, 00:05:48.787 "seek_data": false, 00:05:48.787 "copy": true, 00:05:48.787 "nvme_iov_md": false 00:05:48.787 }, 00:05:48.787 "memory_domains": [ 00:05:48.787 { 00:05:48.787 "dma_device_id": "system", 00:05:48.787 "dma_device_type": 1 00:05:48.787 }, 00:05:48.787 { 00:05:48.787 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:48.787 "dma_device_type": 2 00:05:48.787 } 00:05:48.787 ], 00:05:48.787 "driver_specific": { 00:05:48.787 "passthru": { 00:05:48.787 "name": "Passthru0", 00:05:48.787 "base_bdev_name": "Malloc2" 00:05:48.787 } 00:05:48.787 } 00:05:48.787 } 00:05:48.787 ]' 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:05:48.787 00:05:48.787 real 0m0.321s 00:05:48.787 user 0m0.199s 00:05:48.787 sys 0m0.049s 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:48.787 18:56:06 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.787 ************************************ 00:05:48.787 END TEST rpc_daemon_integrity 00:05:48.787 ************************************ 00:05:49.048 18:56:06 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:05:49.048 18:56:06 rpc -- rpc/rpc.sh@84 -- # killprocess 68836 00:05:49.048 18:56:06 rpc -- common/autotest_common.sh@954 -- # '[' -z 68836 ']' 00:05:49.048 18:56:06 rpc -- common/autotest_common.sh@958 -- # kill -0 68836 00:05:49.048 18:56:06 rpc -- common/autotest_common.sh@959 -- # uname 00:05:49.048 18:56:06 rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:49.048 18:56:06 rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 68836 00:05:49.048 18:56:06 rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:49.048 18:56:06 rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:49.048 killing process with pid 68836 00:05:49.048 18:56:06 rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 68836' 00:05:49.048 18:56:06 rpc -- common/autotest_common.sh@973 -- # kill 68836 00:05:49.048 18:56:06 rpc -- common/autotest_common.sh@978 -- # wait 68836 00:05:49.307 00:05:49.307 real 0m2.800s 00:05:49.307 user 0m3.386s 00:05:49.307 sys 0m0.812s 00:05:49.307 18:56:06 rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:49.307 18:56:06 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:49.307 ************************************ 00:05:49.307 END TEST rpc 00:05:49.307 ************************************ 00:05:49.307 18:56:06 -- spdk/autotest.sh@157 -- # run_test skip_rpc /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:05:49.307 18:56:06 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:49.307 18:56:06 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:49.307 18:56:06 -- common/autotest_common.sh@10 -- # set +x 00:05:49.307 ************************************ 00:05:49.307 START TEST skip_rpc 00:05:49.307 ************************************ 00:05:49.307 18:56:06 skip_rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:05:49.575 * Looking for test storage... 00:05:49.575 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:05:49.575 18:56:06 skip_rpc -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:05:49.575 18:56:06 skip_rpc -- common/autotest_common.sh@1711 -- # lcov --version 00:05:49.575 18:56:06 skip_rpc -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:05:49.575 18:56:07 skip_rpc -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:05:49.575 18:56:07 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@345 -- # : 1 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:49.576 18:56:07 skip_rpc -- scripts/common.sh@368 -- # return 0 00:05:49.576 18:56:07 skip_rpc -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:49.576 18:56:07 skip_rpc -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:05:49.576 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:49.576 --rc genhtml_branch_coverage=1 00:05:49.576 --rc genhtml_function_coverage=1 00:05:49.576 --rc genhtml_legend=1 00:05:49.576 --rc geninfo_all_blocks=1 00:05:49.576 --rc geninfo_unexecuted_blocks=1 00:05:49.576 00:05:49.576 ' 00:05:49.576 18:56:07 skip_rpc -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:05:49.576 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:49.576 --rc genhtml_branch_coverage=1 00:05:49.576 --rc genhtml_function_coverage=1 00:05:49.576 --rc genhtml_legend=1 00:05:49.576 --rc geninfo_all_blocks=1 00:05:49.576 --rc geninfo_unexecuted_blocks=1 00:05:49.576 00:05:49.576 ' 00:05:49.576 18:56:07 skip_rpc -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:05:49.576 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:49.576 --rc genhtml_branch_coverage=1 00:05:49.576 --rc genhtml_function_coverage=1 00:05:49.576 --rc genhtml_legend=1 00:05:49.576 --rc geninfo_all_blocks=1 00:05:49.576 --rc geninfo_unexecuted_blocks=1 00:05:49.576 00:05:49.576 ' 00:05:49.576 18:56:07 skip_rpc -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:05:49.576 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:49.576 --rc genhtml_branch_coverage=1 00:05:49.576 --rc genhtml_function_coverage=1 00:05:49.576 --rc genhtml_legend=1 00:05:49.576 --rc geninfo_all_blocks=1 00:05:49.576 --rc geninfo_unexecuted_blocks=1 00:05:49.576 00:05:49.576 ' 00:05:49.576 18:56:07 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:49.576 18:56:07 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:05:49.576 18:56:07 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:05:49.576 18:56:07 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:49.576 18:56:07 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:49.576 18:56:07 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:49.576 ************************************ 00:05:49.576 START TEST skip_rpc 00:05:49.576 ************************************ 00:05:49.576 18:56:07 skip_rpc.skip_rpc -- common/autotest_common.sh@1129 -- # test_skip_rpc 00:05:49.576 18:56:07 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=69038 00:05:49.576 18:56:07 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:05:49.576 18:56:07 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:49.576 18:56:07 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:05:49.871 [2024-12-05 18:56:07.156125] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:05:49.871 [2024-12-05 18:56:07.156261] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69038 ] 00:05:49.871 [2024-12-05 18:56:07.309367] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:49.871 [2024-12-05 18:56:07.333796] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # local es=0 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd spdk_get_version 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # rpc_cmd spdk_get_version 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # es=1 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 69038 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # '[' -z 69038 ']' 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@958 -- # kill -0 69038 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # uname 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69038 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:55.183 killing process with pid 69038 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69038' 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@973 -- # kill 69038 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@978 -- # wait 69038 00:05:55.183 00:05:55.183 real 0m5.425s 00:05:55.183 user 0m5.042s 00:05:55.183 sys 0m0.314s 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:05:55.183 18:56:12 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:55.183 ************************************ 00:05:55.183 END TEST skip_rpc 00:05:55.183 ************************************ 00:05:55.183 18:56:12 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:05:55.183 18:56:12 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:05:55.183 18:56:12 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:05:55.183 18:56:12 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:55.183 ************************************ 00:05:55.183 START TEST skip_rpc_with_json 00:05:55.183 ************************************ 00:05:55.183 18:56:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_json 00:05:55.183 18:56:12 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:05:55.183 18:56:12 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=69125 00:05:55.183 18:56:12 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:05:55.183 18:56:12 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:55.183 18:56:12 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 69125 00:05:55.183 18:56:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@835 -- # '[' -z 69125 ']' 00:05:55.183 18:56:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:55.183 18:56:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # local max_retries=100 00:05:55.183 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:55.183 18:56:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:55.183 18:56:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@844 -- # xtrace_disable 00:05:55.183 18:56:12 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:55.183 [2024-12-05 18:56:12.648811] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:05:55.183 [2024-12-05 18:56:12.648953] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69125 ] 00:05:55.443 [2024-12-05 18:56:12.803624] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:55.443 [2024-12-05 18:56:12.827942] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:05:56.013 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:05:56.013 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@868 -- # return 0 00:05:56.013 18:56:13 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:05:56.013 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:56.013 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:56.013 [2024-12-05 18:56:13.460883] nvmf_rpc.c:2707:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:05:56.013 request: 00:05:56.013 { 00:05:56.013 "trtype": "tcp", 00:05:56.013 "method": "nvmf_get_transports", 00:05:56.013 "req_id": 1 00:05:56.013 } 00:05:56.013 Got JSON-RPC error response 00:05:56.013 response: 00:05:56.013 { 00:05:56.013 "code": -19, 00:05:56.013 "message": "No such device" 00:05:56.013 } 00:05:56.013 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:05:56.013 18:56:13 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:05:56.013 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:56.013 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:56.013 [2024-12-05 18:56:13.472968] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:05:56.013 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:56.013 18:56:13 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:05:56.013 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:05:56.013 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:56.273 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:05:56.273 18:56:13 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:56.273 { 00:05:56.273 "subsystems": [ 00:05:56.273 { 00:05:56.273 "subsystem": "fsdev", 00:05:56.273 "config": [ 00:05:56.273 { 00:05:56.273 "method": "fsdev_set_opts", 00:05:56.273 "params": { 00:05:56.273 "fsdev_io_pool_size": 65535, 00:05:56.273 "fsdev_io_cache_size": 256 00:05:56.273 } 00:05:56.273 } 00:05:56.273 ] 00:05:56.273 }, 00:05:56.273 { 00:05:56.273 "subsystem": "keyring", 00:05:56.273 "config": [] 00:05:56.273 }, 00:05:56.273 { 00:05:56.273 "subsystem": "iobuf", 00:05:56.273 "config": [ 00:05:56.273 { 00:05:56.273 "method": "iobuf_set_options", 00:05:56.273 "params": { 00:05:56.273 "small_pool_count": 8192, 00:05:56.273 "large_pool_count": 1024, 00:05:56.273 "small_bufsize": 8192, 00:05:56.273 "large_bufsize": 135168, 00:05:56.273 "enable_numa": false 00:05:56.273 } 00:05:56.273 } 00:05:56.273 ] 00:05:56.273 }, 00:05:56.273 { 00:05:56.273 "subsystem": "sock", 00:05:56.273 "config": [ 00:05:56.273 { 00:05:56.273 "method": "sock_set_default_impl", 00:05:56.273 "params": { 00:05:56.273 "impl_name": "posix" 00:05:56.273 } 00:05:56.273 }, 00:05:56.273 { 00:05:56.273 "method": "sock_impl_set_options", 00:05:56.273 "params": { 00:05:56.273 "impl_name": "ssl", 00:05:56.273 "recv_buf_size": 4096, 00:05:56.273 "send_buf_size": 4096, 00:05:56.273 "enable_recv_pipe": true, 00:05:56.273 "enable_quickack": false, 00:05:56.273 "enable_placement_id": 0, 00:05:56.273 "enable_zerocopy_send_server": true, 00:05:56.273 "enable_zerocopy_send_client": false, 00:05:56.273 "zerocopy_threshold": 0, 00:05:56.273 "tls_version": 0, 00:05:56.273 "enable_ktls": false 00:05:56.273 } 00:05:56.273 }, 00:05:56.273 { 00:05:56.273 "method": "sock_impl_set_options", 00:05:56.273 "params": { 00:05:56.273 "impl_name": "posix", 00:05:56.273 "recv_buf_size": 2097152, 00:05:56.273 "send_buf_size": 2097152, 00:05:56.273 "enable_recv_pipe": true, 00:05:56.273 "enable_quickack": false, 00:05:56.273 "enable_placement_id": 0, 00:05:56.273 "enable_zerocopy_send_server": true, 00:05:56.273 "enable_zerocopy_send_client": false, 00:05:56.273 "zerocopy_threshold": 0, 00:05:56.273 "tls_version": 0, 00:05:56.273 "enable_ktls": false 00:05:56.273 } 00:05:56.273 } 00:05:56.273 ] 00:05:56.273 }, 00:05:56.273 { 00:05:56.273 "subsystem": "vmd", 00:05:56.273 "config": [] 00:05:56.273 }, 00:05:56.273 { 00:05:56.273 "subsystem": "accel", 00:05:56.273 "config": [ 00:05:56.273 { 00:05:56.273 "method": "accel_set_options", 00:05:56.273 "params": { 00:05:56.273 "small_cache_size": 128, 00:05:56.273 "large_cache_size": 16, 00:05:56.273 "task_count": 2048, 00:05:56.273 "sequence_count": 2048, 00:05:56.273 "buf_count": 2048 00:05:56.273 } 00:05:56.273 } 00:05:56.273 ] 00:05:56.273 }, 00:05:56.273 { 00:05:56.273 "subsystem": "bdev", 00:05:56.273 "config": [ 00:05:56.273 { 00:05:56.273 "method": "bdev_set_options", 00:05:56.273 "params": { 00:05:56.273 "bdev_io_pool_size": 65535, 00:05:56.273 "bdev_io_cache_size": 256, 00:05:56.274 "bdev_auto_examine": true, 00:05:56.274 "iobuf_small_cache_size": 128, 00:05:56.274 "iobuf_large_cache_size": 16 00:05:56.274 } 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "method": "bdev_raid_set_options", 00:05:56.274 "params": { 00:05:56.274 "process_window_size_kb": 1024, 00:05:56.274 "process_max_bandwidth_mb_sec": 0 00:05:56.274 } 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "method": "bdev_iscsi_set_options", 00:05:56.274 "params": { 00:05:56.274 "timeout_sec": 30 00:05:56.274 } 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "method": "bdev_nvme_set_options", 00:05:56.274 "params": { 00:05:56.274 "action_on_timeout": "none", 00:05:56.274 "timeout_us": 0, 00:05:56.274 "timeout_admin_us": 0, 00:05:56.274 "keep_alive_timeout_ms": 10000, 00:05:56.274 "arbitration_burst": 0, 00:05:56.274 "low_priority_weight": 0, 00:05:56.274 "medium_priority_weight": 0, 00:05:56.274 "high_priority_weight": 0, 00:05:56.274 "nvme_adminq_poll_period_us": 10000, 00:05:56.274 "nvme_ioq_poll_period_us": 0, 00:05:56.274 "io_queue_requests": 0, 00:05:56.274 "delay_cmd_submit": true, 00:05:56.274 "transport_retry_count": 4, 00:05:56.274 "bdev_retry_count": 3, 00:05:56.274 "transport_ack_timeout": 0, 00:05:56.274 "ctrlr_loss_timeout_sec": 0, 00:05:56.274 "reconnect_delay_sec": 0, 00:05:56.274 "fast_io_fail_timeout_sec": 0, 00:05:56.274 "disable_auto_failback": false, 00:05:56.274 "generate_uuids": false, 00:05:56.274 "transport_tos": 0, 00:05:56.274 "nvme_error_stat": false, 00:05:56.274 "rdma_srq_size": 0, 00:05:56.274 "io_path_stat": false, 00:05:56.274 "allow_accel_sequence": false, 00:05:56.274 "rdma_max_cq_size": 0, 00:05:56.274 "rdma_cm_event_timeout_ms": 0, 00:05:56.274 "dhchap_digests": [ 00:05:56.274 "sha256", 00:05:56.274 "sha384", 00:05:56.274 "sha512" 00:05:56.274 ], 00:05:56.274 "dhchap_dhgroups": [ 00:05:56.274 "null", 00:05:56.274 "ffdhe2048", 00:05:56.274 "ffdhe3072", 00:05:56.274 "ffdhe4096", 00:05:56.274 "ffdhe6144", 00:05:56.274 "ffdhe8192" 00:05:56.274 ] 00:05:56.274 } 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "method": "bdev_nvme_set_hotplug", 00:05:56.274 "params": { 00:05:56.274 "period_us": 100000, 00:05:56.274 "enable": false 00:05:56.274 } 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "method": "bdev_wait_for_examine" 00:05:56.274 } 00:05:56.274 ] 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "subsystem": "scsi", 00:05:56.274 "config": null 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "subsystem": "scheduler", 00:05:56.274 "config": [ 00:05:56.274 { 00:05:56.274 "method": "framework_set_scheduler", 00:05:56.274 "params": { 00:05:56.274 "name": "static" 00:05:56.274 } 00:05:56.274 } 00:05:56.274 ] 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "subsystem": "vhost_scsi", 00:05:56.274 "config": [] 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "subsystem": "vhost_blk", 00:05:56.274 "config": [] 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "subsystem": "ublk", 00:05:56.274 "config": [] 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "subsystem": "nbd", 00:05:56.274 "config": [] 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "subsystem": "nvmf", 00:05:56.274 "config": [ 00:05:56.274 { 00:05:56.274 "method": "nvmf_set_config", 00:05:56.274 "params": { 00:05:56.274 "discovery_filter": "match_any", 00:05:56.274 "admin_cmd_passthru": { 00:05:56.274 "identify_ctrlr": false 00:05:56.274 }, 00:05:56.274 "dhchap_digests": [ 00:05:56.274 "sha256", 00:05:56.274 "sha384", 00:05:56.274 "sha512" 00:05:56.274 ], 00:05:56.274 "dhchap_dhgroups": [ 00:05:56.274 "null", 00:05:56.274 "ffdhe2048", 00:05:56.274 "ffdhe3072", 00:05:56.274 "ffdhe4096", 00:05:56.274 "ffdhe6144", 00:05:56.274 "ffdhe8192" 00:05:56.274 ] 00:05:56.274 } 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "method": "nvmf_set_max_subsystems", 00:05:56.274 "params": { 00:05:56.274 "max_subsystems": 1024 00:05:56.274 } 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "method": "nvmf_set_crdt", 00:05:56.274 "params": { 00:05:56.274 "crdt1": 0, 00:05:56.274 "crdt2": 0, 00:05:56.274 "crdt3": 0 00:05:56.274 } 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "method": "nvmf_create_transport", 00:05:56.274 "params": { 00:05:56.274 "trtype": "TCP", 00:05:56.274 "max_queue_depth": 128, 00:05:56.274 "max_io_qpairs_per_ctrlr": 127, 00:05:56.274 "in_capsule_data_size": 4096, 00:05:56.274 "max_io_size": 131072, 00:05:56.274 "io_unit_size": 131072, 00:05:56.274 "max_aq_depth": 128, 00:05:56.274 "num_shared_buffers": 511, 00:05:56.274 "buf_cache_size": 4294967295, 00:05:56.274 "dif_insert_or_strip": false, 00:05:56.274 "zcopy": false, 00:05:56.274 "c2h_success": true, 00:05:56.274 "sock_priority": 0, 00:05:56.274 "abort_timeout_sec": 1, 00:05:56.274 "ack_timeout": 0, 00:05:56.274 "data_wr_pool_size": 0 00:05:56.274 } 00:05:56.274 } 00:05:56.274 ] 00:05:56.274 }, 00:05:56.274 { 00:05:56.274 "subsystem": "iscsi", 00:05:56.274 "config": [ 00:05:56.274 { 00:05:56.274 "method": "iscsi_set_options", 00:05:56.274 "params": { 00:05:56.274 "node_base": "iqn.2016-06.io.spdk", 00:05:56.274 "max_sessions": 128, 00:05:56.274 "max_connections_per_session": 2, 00:05:56.274 "max_queue_depth": 64, 00:05:56.274 "default_time2wait": 2, 00:05:56.274 "default_time2retain": 20, 00:05:56.274 "first_burst_length": 8192, 00:05:56.274 "immediate_data": true, 00:05:56.274 "allow_duplicated_isid": false, 00:05:56.274 "error_recovery_level": 0, 00:05:56.274 "nop_timeout": 60, 00:05:56.274 "nop_in_interval": 30, 00:05:56.274 "disable_chap": false, 00:05:56.274 "require_chap": false, 00:05:56.274 "mutual_chap": false, 00:05:56.274 "chap_group": 0, 00:05:56.274 "max_large_datain_per_connection": 64, 00:05:56.274 "max_r2t_per_connection": 4, 00:05:56.274 "pdu_pool_size": 36864, 00:05:56.275 "immediate_data_pool_size": 16384, 00:05:56.275 "data_out_pool_size": 2048 00:05:56.275 } 00:05:56.275 } 00:05:56.275 ] 00:05:56.275 } 00:05:56.275 ] 00:05:56.275 } 00:05:56.275 18:56:13 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:05:56.275 18:56:13 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 69125 00:05:56.275 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 69125 ']' 00:05:56.275 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 69125 00:05:56.275 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:05:56.275 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:05:56.275 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69125 00:05:56.275 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:05:56.275 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:05:56.275 killing process with pid 69125 00:05:56.275 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69125' 00:05:56.275 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 69125 00:05:56.275 18:56:13 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 69125 00:05:56.535 18:56:14 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:56.535 18:56:14 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=69154 00:05:56.535 18:56:14 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:06:01.816 18:56:19 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 69154 00:06:01.816 18:56:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 69154 ']' 00:06:01.816 18:56:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 69154 00:06:01.816 18:56:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:06:01.816 18:56:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:01.816 18:56:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69154 00:06:01.816 18:56:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:01.816 18:56:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:01.816 killing process with pid 69154 00:06:01.816 18:56:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69154' 00:06:01.816 18:56:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 69154 00:06:01.816 18:56:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 69154 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:06:02.076 00:06:02.076 real 0m6.916s 00:06:02.076 user 0m6.484s 00:06:02.076 sys 0m0.707s 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:02.076 ************************************ 00:06:02.076 END TEST skip_rpc_with_json 00:06:02.076 ************************************ 00:06:02.076 18:56:19 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:06:02.076 18:56:19 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:02.076 18:56:19 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:02.076 18:56:19 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:02.076 ************************************ 00:06:02.076 START TEST skip_rpc_with_delay 00:06:02.076 ************************************ 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_delay 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # local es=0 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:06:02.076 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:02.336 [2024-12-05 18:56:19.636976] app.c: 842:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:06:02.336 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # es=1 00:06:02.336 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:02.336 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:02.336 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:02.336 00:06:02.336 real 0m0.157s 00:06:02.336 user 0m0.086s 00:06:02.336 sys 0m0.070s 00:06:02.336 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:02.336 18:56:19 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:06:02.336 ************************************ 00:06:02.336 END TEST skip_rpc_with_delay 00:06:02.336 ************************************ 00:06:02.336 18:56:19 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:06:02.336 18:56:19 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:06:02.336 18:56:19 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:06:02.336 18:56:19 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:02.336 18:56:19 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:02.336 18:56:19 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:02.336 ************************************ 00:06:02.336 START TEST exit_on_failed_rpc_init 00:06:02.336 ************************************ 00:06:02.336 18:56:19 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1129 -- # test_exit_on_failed_rpc_init 00:06:02.336 18:56:19 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=69262 00:06:02.336 18:56:19 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:02.336 18:56:19 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 69262 00:06:02.336 18:56:19 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@835 -- # '[' -z 69262 ']' 00:06:02.336 18:56:19 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:02.336 18:56:19 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:02.336 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:02.336 18:56:19 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:02.336 18:56:19 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:02.336 18:56:19 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:06:02.336 [2024-12-05 18:56:19.864853] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:02.336 [2024-12-05 18:56:19.865001] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69262 ] 00:06:02.596 [2024-12-05 18:56:20.016999] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:02.596 [2024-12-05 18:56:20.041715] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:03.166 18:56:20 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:03.166 18:56:20 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@868 -- # return 0 00:06:03.166 18:56:20 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:03.166 18:56:20 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:03.166 18:56:20 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # local es=0 00:06:03.166 18:56:20 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:03.166 18:56:20 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:03.166 18:56:20 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:03.166 18:56:20 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:03.166 18:56:20 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:03.166 18:56:20 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:03.166 18:56:20 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:03.166 18:56:20 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:03.166 18:56:20 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:06:03.166 18:56:20 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:03.426 [2024-12-05 18:56:20.764446] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:03.426 [2024-12-05 18:56:20.764587] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69278 ] 00:06:03.426 [2024-12-05 18:56:20.915839] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:03.426 [2024-12-05 18:56:20.942769] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:03.426 [2024-12-05 18:56:20.942853] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:06:03.426 [2024-12-05 18:56:20.942879] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:06:03.426 [2024-12-05 18:56:20.942897] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # es=234 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@664 -- # es=106 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@665 -- # case "$es" in 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@672 -- # es=1 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 69262 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # '[' -z 69262 ']' 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@958 -- # kill -0 69262 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # uname 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69262 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:03.686 killing process with pid 69262 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69262' 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@973 -- # kill 69262 00:06:03.686 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@978 -- # wait 69262 00:06:03.946 00:06:03.946 real 0m1.657s 00:06:03.946 user 0m1.745s 00:06:03.946 sys 0m0.480s 00:06:03.946 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:03.946 18:56:21 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:06:03.946 ************************************ 00:06:03.946 END TEST exit_on_failed_rpc_init 00:06:03.946 ************************************ 00:06:03.946 18:56:21 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:06:03.946 00:06:03.946 real 0m14.659s 00:06:03.946 user 0m13.557s 00:06:03.946 sys 0m1.885s 00:06:03.946 18:56:21 skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:03.946 18:56:21 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:03.946 ************************************ 00:06:03.946 END TEST skip_rpc 00:06:03.946 ************************************ 00:06:04.204 18:56:21 -- spdk/autotest.sh@158 -- # run_test rpc_client /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:06:04.204 18:56:21 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:04.204 18:56:21 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:04.204 18:56:21 -- common/autotest_common.sh@10 -- # set +x 00:06:04.204 ************************************ 00:06:04.205 START TEST rpc_client 00:06:04.205 ************************************ 00:06:04.205 18:56:21 rpc_client -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:06:04.205 * Looking for test storage... 00:06:04.205 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc_client 00:06:04.205 18:56:21 rpc_client -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:06:04.205 18:56:21 rpc_client -- common/autotest_common.sh@1711 -- # lcov --version 00:06:04.205 18:56:21 rpc_client -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:06:04.463 18:56:21 rpc_client -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@345 -- # : 1 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@353 -- # local d=1 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@355 -- # echo 1 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@353 -- # local d=2 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@355 -- # echo 2 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:04.463 18:56:21 rpc_client -- scripts/common.sh@368 -- # return 0 00:06:04.463 18:56:21 rpc_client -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:04.463 18:56:21 rpc_client -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:06:04.464 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:04.464 --rc genhtml_branch_coverage=1 00:06:04.464 --rc genhtml_function_coverage=1 00:06:04.464 --rc genhtml_legend=1 00:06:04.464 --rc geninfo_all_blocks=1 00:06:04.464 --rc geninfo_unexecuted_blocks=1 00:06:04.464 00:06:04.464 ' 00:06:04.464 18:56:21 rpc_client -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:06:04.464 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:04.464 --rc genhtml_branch_coverage=1 00:06:04.464 --rc genhtml_function_coverage=1 00:06:04.464 --rc genhtml_legend=1 00:06:04.464 --rc geninfo_all_blocks=1 00:06:04.464 --rc geninfo_unexecuted_blocks=1 00:06:04.464 00:06:04.464 ' 00:06:04.464 18:56:21 rpc_client -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:06:04.464 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:04.464 --rc genhtml_branch_coverage=1 00:06:04.464 --rc genhtml_function_coverage=1 00:06:04.464 --rc genhtml_legend=1 00:06:04.464 --rc geninfo_all_blocks=1 00:06:04.464 --rc geninfo_unexecuted_blocks=1 00:06:04.464 00:06:04.464 ' 00:06:04.464 18:56:21 rpc_client -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:06:04.464 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:04.464 --rc genhtml_branch_coverage=1 00:06:04.464 --rc genhtml_function_coverage=1 00:06:04.464 --rc genhtml_legend=1 00:06:04.464 --rc geninfo_all_blocks=1 00:06:04.464 --rc geninfo_unexecuted_blocks=1 00:06:04.464 00:06:04.464 ' 00:06:04.464 18:56:21 rpc_client -- rpc_client/rpc_client.sh@10 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client_test 00:06:04.464 OK 00:06:04.464 18:56:21 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:06:04.464 00:06:04.464 real 0m0.285s 00:06:04.464 user 0m0.150s 00:06:04.464 sys 0m0.155s 00:06:04.464 18:56:21 rpc_client -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:04.464 18:56:21 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:06:04.464 ************************************ 00:06:04.464 END TEST rpc_client 00:06:04.464 ************************************ 00:06:04.464 18:56:21 -- spdk/autotest.sh@159 -- # run_test json_config /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:06:04.464 18:56:21 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:04.464 18:56:21 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:04.464 18:56:21 -- common/autotest_common.sh@10 -- # set +x 00:06:04.464 ************************************ 00:06:04.464 START TEST json_config 00:06:04.464 ************************************ 00:06:04.464 18:56:21 json_config -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:06:04.464 18:56:22 json_config -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:06:04.464 18:56:22 json_config -- common/autotest_common.sh@1711 -- # lcov --version 00:06:04.464 18:56:22 json_config -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:06:04.724 18:56:22 json_config -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:06:04.724 18:56:22 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:04.724 18:56:22 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:04.724 18:56:22 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:04.724 18:56:22 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:06:04.724 18:56:22 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:06:04.724 18:56:22 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:06:04.724 18:56:22 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:06:04.724 18:56:22 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:06:04.724 18:56:22 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:06:04.724 18:56:22 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:06:04.724 18:56:22 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:04.724 18:56:22 json_config -- scripts/common.sh@344 -- # case "$op" in 00:06:04.724 18:56:22 json_config -- scripts/common.sh@345 -- # : 1 00:06:04.724 18:56:22 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:04.724 18:56:22 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:04.724 18:56:22 json_config -- scripts/common.sh@365 -- # decimal 1 00:06:04.724 18:56:22 json_config -- scripts/common.sh@353 -- # local d=1 00:06:04.724 18:56:22 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:04.724 18:56:22 json_config -- scripts/common.sh@355 -- # echo 1 00:06:04.724 18:56:22 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:06:04.724 18:56:22 json_config -- scripts/common.sh@366 -- # decimal 2 00:06:04.724 18:56:22 json_config -- scripts/common.sh@353 -- # local d=2 00:06:04.724 18:56:22 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:04.724 18:56:22 json_config -- scripts/common.sh@355 -- # echo 2 00:06:04.724 18:56:22 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:06:04.724 18:56:22 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:04.724 18:56:22 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:04.724 18:56:22 json_config -- scripts/common.sh@368 -- # return 0 00:06:04.724 18:56:22 json_config -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:04.724 18:56:22 json_config -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:06:04.724 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:04.724 --rc genhtml_branch_coverage=1 00:06:04.724 --rc genhtml_function_coverage=1 00:06:04.724 --rc genhtml_legend=1 00:06:04.724 --rc geninfo_all_blocks=1 00:06:04.724 --rc geninfo_unexecuted_blocks=1 00:06:04.724 00:06:04.724 ' 00:06:04.724 18:56:22 json_config -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:06:04.724 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:04.724 --rc genhtml_branch_coverage=1 00:06:04.724 --rc genhtml_function_coverage=1 00:06:04.724 --rc genhtml_legend=1 00:06:04.724 --rc geninfo_all_blocks=1 00:06:04.725 --rc geninfo_unexecuted_blocks=1 00:06:04.725 00:06:04.725 ' 00:06:04.725 18:56:22 json_config -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:06:04.725 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:04.725 --rc genhtml_branch_coverage=1 00:06:04.725 --rc genhtml_function_coverage=1 00:06:04.725 --rc genhtml_legend=1 00:06:04.725 --rc geninfo_all_blocks=1 00:06:04.725 --rc geninfo_unexecuted_blocks=1 00:06:04.725 00:06:04.725 ' 00:06:04.725 18:56:22 json_config -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:06:04.725 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:04.725 --rc genhtml_branch_coverage=1 00:06:04.725 --rc genhtml_function_coverage=1 00:06:04.725 --rc genhtml_legend=1 00:06:04.725 --rc geninfo_all_blocks=1 00:06:04.725 --rc geninfo_unexecuted_blocks=1 00:06:04.725 00:06:04.725 ' 00:06:04.725 18:56:22 json_config -- json_config/json_config.sh@8 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@7 -- # uname -s 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:4da6608a-058c-4cf5-8acb-b90b56f362fe 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=4da6608a-058c-4cf5-8acb-b90b56f362fe 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:06:04.725 18:56:22 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:06:04.725 18:56:22 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:04.725 18:56:22 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:04.725 18:56:22 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:04.725 18:56:22 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:04.725 18:56:22 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:04.725 18:56:22 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:04.725 18:56:22 json_config -- paths/export.sh@5 -- # export PATH 00:06:04.725 18:56:22 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@51 -- # : 0 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:04.725 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:04.725 18:56:22 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:04.725 18:56:22 json_config -- json_config/json_config.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:06:04.725 18:56:22 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:06:04.725 18:56:22 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:06:04.725 18:56:22 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:06:04.725 18:56:22 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:06:04.725 WARNING: No tests are enabled so not running JSON configuration tests 00:06:04.725 18:56:22 json_config -- json_config/json_config.sh@27 -- # echo 'WARNING: No tests are enabled so not running JSON configuration tests' 00:06:04.725 18:56:22 json_config -- json_config/json_config.sh@28 -- # exit 0 00:06:04.725 00:06:04.725 real 0m0.228s 00:06:04.725 user 0m0.135s 00:06:04.725 sys 0m0.102s 00:06:04.725 18:56:22 json_config -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:04.725 18:56:22 json_config -- common/autotest_common.sh@10 -- # set +x 00:06:04.725 ************************************ 00:06:04.725 END TEST json_config 00:06:04.725 ************************************ 00:06:04.725 18:56:22 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:06:04.725 18:56:22 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:04.725 18:56:22 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:04.725 18:56:22 -- common/autotest_common.sh@10 -- # set +x 00:06:04.725 ************************************ 00:06:04.725 START TEST json_config_extra_key 00:06:04.725 ************************************ 00:06:04.725 18:56:22 json_config_extra_key -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:06:04.986 18:56:22 json_config_extra_key -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:06:04.986 18:56:22 json_config_extra_key -- common/autotest_common.sh@1711 -- # lcov --version 00:06:04.986 18:56:22 json_config_extra_key -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:06:04.986 18:56:22 json_config_extra_key -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:04.986 18:56:22 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:06:04.987 18:56:22 json_config_extra_key -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:04.987 18:56:22 json_config_extra_key -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:06:04.987 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:04.987 --rc genhtml_branch_coverage=1 00:06:04.987 --rc genhtml_function_coverage=1 00:06:04.987 --rc genhtml_legend=1 00:06:04.987 --rc geninfo_all_blocks=1 00:06:04.987 --rc geninfo_unexecuted_blocks=1 00:06:04.987 00:06:04.987 ' 00:06:04.987 18:56:22 json_config_extra_key -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:06:04.987 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:04.987 --rc genhtml_branch_coverage=1 00:06:04.987 --rc genhtml_function_coverage=1 00:06:04.987 --rc genhtml_legend=1 00:06:04.987 --rc geninfo_all_blocks=1 00:06:04.987 --rc geninfo_unexecuted_blocks=1 00:06:04.987 00:06:04.987 ' 00:06:04.987 18:56:22 json_config_extra_key -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:06:04.987 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:04.987 --rc genhtml_branch_coverage=1 00:06:04.987 --rc genhtml_function_coverage=1 00:06:04.987 --rc genhtml_legend=1 00:06:04.987 --rc geninfo_all_blocks=1 00:06:04.987 --rc geninfo_unexecuted_blocks=1 00:06:04.987 00:06:04.987 ' 00:06:04.987 18:56:22 json_config_extra_key -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:06:04.987 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:04.987 --rc genhtml_branch_coverage=1 00:06:04.987 --rc genhtml_function_coverage=1 00:06:04.987 --rc genhtml_legend=1 00:06:04.987 --rc geninfo_all_blocks=1 00:06:04.987 --rc geninfo_unexecuted_blocks=1 00:06:04.987 00:06:04.987 ' 00:06:04.987 18:56:22 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:4da6608a-058c-4cf5-8acb-b90b56f362fe 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=4da6608a-058c-4cf5-8acb-b90b56f362fe 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:06:04.987 18:56:22 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:06:04.987 18:56:22 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:04.987 18:56:22 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:04.987 18:56:22 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:04.987 18:56:22 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:04.987 18:56:22 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:04.987 18:56:22 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:04.987 18:56:22 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:06:04.987 18:56:22 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:04.987 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:04.987 18:56:22 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:04.987 18:56:22 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:06:04.987 18:56:22 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:06:04.987 18:56:22 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:06:04.987 18:56:22 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:06:04.987 18:56:22 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:06:04.987 18:56:22 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:06:04.987 18:56:22 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:06:04.987 18:56:22 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json') 00:06:04.987 18:56:22 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:06:04.987 18:56:22 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:06:04.987 18:56:22 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:06:04.987 INFO: launching applications... 00:06:04.987 18:56:22 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:06:04.987 18:56:22 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:06:04.987 18:56:22 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:06:04.987 18:56:22 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:06:04.987 18:56:22 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:06:04.987 18:56:22 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:06:04.987 18:56:22 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:06:04.987 18:56:22 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:06:04.987 18:56:22 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=69466 00:06:04.987 18:56:22 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:06:04.987 Waiting for target to run... 00:06:04.987 18:56:22 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 69466 /var/tmp/spdk_tgt.sock 00:06:04.987 18:56:22 json_config_extra_key -- json_config/common.sh@21 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:06:04.987 18:56:22 json_config_extra_key -- common/autotest_common.sh@835 -- # '[' -z 69466 ']' 00:06:04.987 18:56:22 json_config_extra_key -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:06:04.987 18:56:22 json_config_extra_key -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:04.987 18:56:22 json_config_extra_key -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:06:04.987 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:06:04.987 18:56:22 json_config_extra_key -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:04.987 18:56:22 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:06:04.987 [2024-12-05 18:56:22.523182] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:04.987 [2024-12-05 18:56:22.523391] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69466 ] 00:06:05.556 [2024-12-05 18:56:22.886917] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:05.556 [2024-12-05 18:56:22.902007] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:05.816 18:56:23 json_config_extra_key -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:05.816 00:06:05.816 INFO: shutting down applications... 00:06:05.816 18:56:23 json_config_extra_key -- common/autotest_common.sh@868 -- # return 0 00:06:05.816 18:56:23 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:06:05.816 18:56:23 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:06:05.816 18:56:23 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:06:05.816 18:56:23 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:06:05.816 18:56:23 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:06:05.816 18:56:23 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 69466 ]] 00:06:05.816 18:56:23 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 69466 00:06:05.816 18:56:23 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:06:05.816 18:56:23 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:06:05.816 18:56:23 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69466 00:06:05.816 18:56:23 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:06:06.385 18:56:23 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:06:06.385 18:56:23 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:06:06.385 18:56:23 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69466 00:06:06.385 18:56:23 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:06:06.385 18:56:23 json_config_extra_key -- json_config/common.sh@43 -- # break 00:06:06.385 18:56:23 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:06:06.385 18:56:23 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:06:06.385 SPDK target shutdown done 00:06:06.385 18:56:23 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:06:06.385 Success 00:06:06.385 00:06:06.385 real 0m1.649s 00:06:06.385 user 0m1.346s 00:06:06.385 sys 0m0.471s 00:06:06.385 18:56:23 json_config_extra_key -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:06.385 18:56:23 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:06:06.385 ************************************ 00:06:06.385 END TEST json_config_extra_key 00:06:06.386 ************************************ 00:06:06.386 18:56:23 -- spdk/autotest.sh@161 -- # run_test alias_rpc /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:06:06.386 18:56:23 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:06.386 18:56:23 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:06.386 18:56:23 -- common/autotest_common.sh@10 -- # set +x 00:06:06.386 ************************************ 00:06:06.386 START TEST alias_rpc 00:06:06.386 ************************************ 00:06:06.386 18:56:23 alias_rpc -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:06:06.645 * Looking for test storage... 00:06:06.645 * Found test storage at /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc 00:06:06.645 18:56:24 alias_rpc -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:06:06.646 18:56:24 alias_rpc -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:06:06.646 18:56:24 alias_rpc -- common/autotest_common.sh@1711 -- # lcov --version 00:06:06.646 18:56:24 alias_rpc -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@345 -- # : 1 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:06.646 18:56:24 alias_rpc -- scripts/common.sh@368 -- # return 0 00:06:06.646 18:56:24 alias_rpc -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:06.646 18:56:24 alias_rpc -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:06:06.646 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:06.646 --rc genhtml_branch_coverage=1 00:06:06.646 --rc genhtml_function_coverage=1 00:06:06.646 --rc genhtml_legend=1 00:06:06.646 --rc geninfo_all_blocks=1 00:06:06.646 --rc geninfo_unexecuted_blocks=1 00:06:06.646 00:06:06.646 ' 00:06:06.646 18:56:24 alias_rpc -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:06:06.646 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:06.646 --rc genhtml_branch_coverage=1 00:06:06.646 --rc genhtml_function_coverage=1 00:06:06.646 --rc genhtml_legend=1 00:06:06.646 --rc geninfo_all_blocks=1 00:06:06.646 --rc geninfo_unexecuted_blocks=1 00:06:06.646 00:06:06.646 ' 00:06:06.646 18:56:24 alias_rpc -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:06:06.646 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:06.646 --rc genhtml_branch_coverage=1 00:06:06.646 --rc genhtml_function_coverage=1 00:06:06.646 --rc genhtml_legend=1 00:06:06.646 --rc geninfo_all_blocks=1 00:06:06.646 --rc geninfo_unexecuted_blocks=1 00:06:06.646 00:06:06.646 ' 00:06:06.646 18:56:24 alias_rpc -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:06:06.646 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:06.646 --rc genhtml_branch_coverage=1 00:06:06.646 --rc genhtml_function_coverage=1 00:06:06.646 --rc genhtml_legend=1 00:06:06.646 --rc geninfo_all_blocks=1 00:06:06.646 --rc geninfo_unexecuted_blocks=1 00:06:06.646 00:06:06.646 ' 00:06:06.646 18:56:24 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:06:06.646 18:56:24 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=69545 00:06:06.646 18:56:24 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:06.646 18:56:24 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 69545 00:06:06.646 18:56:24 alias_rpc -- common/autotest_common.sh@835 -- # '[' -z 69545 ']' 00:06:06.646 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:06.646 18:56:24 alias_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:06.646 18:56:24 alias_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:06.646 18:56:24 alias_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:06.646 18:56:24 alias_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:06.646 18:56:24 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:06.906 [2024-12-05 18:56:24.223283] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:06.906 [2024-12-05 18:56:24.223440] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69545 ] 00:06:06.906 [2024-12-05 18:56:24.380133] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:06.906 [2024-12-05 18:56:24.405444] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:07.852 18:56:25 alias_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:07.852 18:56:25 alias_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:07.852 18:56:25 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py load_config -i 00:06:07.852 18:56:25 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 69545 00:06:07.852 18:56:25 alias_rpc -- common/autotest_common.sh@954 -- # '[' -z 69545 ']' 00:06:07.852 18:56:25 alias_rpc -- common/autotest_common.sh@958 -- # kill -0 69545 00:06:07.852 18:56:25 alias_rpc -- common/autotest_common.sh@959 -- # uname 00:06:07.852 18:56:25 alias_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:07.852 18:56:25 alias_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69545 00:06:07.852 killing process with pid 69545 00:06:07.852 18:56:25 alias_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:07.852 18:56:25 alias_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:07.852 18:56:25 alias_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69545' 00:06:07.852 18:56:25 alias_rpc -- common/autotest_common.sh@973 -- # kill 69545 00:06:07.852 18:56:25 alias_rpc -- common/autotest_common.sh@978 -- # wait 69545 00:06:08.148 ************************************ 00:06:08.148 END TEST alias_rpc 00:06:08.148 ************************************ 00:06:08.148 00:06:08.148 real 0m1.726s 00:06:08.148 user 0m1.739s 00:06:08.148 sys 0m0.484s 00:06:08.148 18:56:25 alias_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:08.148 18:56:25 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:08.148 18:56:25 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:06:08.148 18:56:25 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:06:08.148 18:56:25 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:08.148 18:56:25 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:08.148 18:56:25 -- common/autotest_common.sh@10 -- # set +x 00:06:08.407 ************************************ 00:06:08.407 START TEST spdkcli_tcp 00:06:08.407 ************************************ 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:06:08.407 * Looking for test storage... 00:06:08.407 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@1711 -- # lcov --version 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:08.407 18:56:25 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:06:08.407 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:08.407 --rc genhtml_branch_coverage=1 00:06:08.407 --rc genhtml_function_coverage=1 00:06:08.407 --rc genhtml_legend=1 00:06:08.407 --rc geninfo_all_blocks=1 00:06:08.407 --rc geninfo_unexecuted_blocks=1 00:06:08.407 00:06:08.407 ' 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:06:08.407 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:08.407 --rc genhtml_branch_coverage=1 00:06:08.407 --rc genhtml_function_coverage=1 00:06:08.407 --rc genhtml_legend=1 00:06:08.407 --rc geninfo_all_blocks=1 00:06:08.407 --rc geninfo_unexecuted_blocks=1 00:06:08.407 00:06:08.407 ' 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:06:08.407 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:08.407 --rc genhtml_branch_coverage=1 00:06:08.407 --rc genhtml_function_coverage=1 00:06:08.407 --rc genhtml_legend=1 00:06:08.407 --rc geninfo_all_blocks=1 00:06:08.407 --rc geninfo_unexecuted_blocks=1 00:06:08.407 00:06:08.407 ' 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:06:08.407 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:08.407 --rc genhtml_branch_coverage=1 00:06:08.407 --rc genhtml_function_coverage=1 00:06:08.407 --rc genhtml_legend=1 00:06:08.407 --rc geninfo_all_blocks=1 00:06:08.407 --rc geninfo_unexecuted_blocks=1 00:06:08.407 00:06:08.407 ' 00:06:08.407 18:56:25 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:06:08.407 18:56:25 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:06:08.407 18:56:25 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:06:08.407 18:56:25 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:06:08.407 18:56:25 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:06:08.407 18:56:25 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:06:08.407 18:56:25 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:08.407 18:56:25 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=69619 00:06:08.407 18:56:25 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:06:08.407 18:56:25 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 69619 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@835 -- # '[' -z 69619 ']' 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:08.407 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:08.407 18:56:25 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:08.666 [2024-12-05 18:56:26.034121] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:08.666 [2024-12-05 18:56:26.034849] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69619 ] 00:06:08.666 [2024-12-05 18:56:26.189585] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:08.666 [2024-12-05 18:56:26.215202] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:08.666 [2024-12-05 18:56:26.215293] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:09.603 18:56:26 spdkcli_tcp -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:09.603 18:56:26 spdkcli_tcp -- common/autotest_common.sh@868 -- # return 0 00:06:09.603 18:56:26 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:06:09.603 18:56:26 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=69636 00:06:09.604 18:56:26 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:06:09.604 [ 00:06:09.604 "bdev_malloc_delete", 00:06:09.604 "bdev_malloc_create", 00:06:09.604 "bdev_null_resize", 00:06:09.604 "bdev_null_delete", 00:06:09.604 "bdev_null_create", 00:06:09.604 "bdev_nvme_cuse_unregister", 00:06:09.604 "bdev_nvme_cuse_register", 00:06:09.604 "bdev_opal_new_user", 00:06:09.604 "bdev_opal_set_lock_state", 00:06:09.604 "bdev_opal_delete", 00:06:09.604 "bdev_opal_get_info", 00:06:09.604 "bdev_opal_create", 00:06:09.604 "bdev_nvme_opal_revert", 00:06:09.604 "bdev_nvme_opal_init", 00:06:09.604 "bdev_nvme_send_cmd", 00:06:09.604 "bdev_nvme_set_keys", 00:06:09.604 "bdev_nvme_get_path_iostat", 00:06:09.604 "bdev_nvme_get_mdns_discovery_info", 00:06:09.604 "bdev_nvme_stop_mdns_discovery", 00:06:09.604 "bdev_nvme_start_mdns_discovery", 00:06:09.604 "bdev_nvme_set_multipath_policy", 00:06:09.604 "bdev_nvme_set_preferred_path", 00:06:09.604 "bdev_nvme_get_io_paths", 00:06:09.604 "bdev_nvme_remove_error_injection", 00:06:09.604 "bdev_nvme_add_error_injection", 00:06:09.604 "bdev_nvme_get_discovery_info", 00:06:09.604 "bdev_nvme_stop_discovery", 00:06:09.604 "bdev_nvme_start_discovery", 00:06:09.604 "bdev_nvme_get_controller_health_info", 00:06:09.604 "bdev_nvme_disable_controller", 00:06:09.604 "bdev_nvme_enable_controller", 00:06:09.604 "bdev_nvme_reset_controller", 00:06:09.604 "bdev_nvme_get_transport_statistics", 00:06:09.604 "bdev_nvme_apply_firmware", 00:06:09.604 "bdev_nvme_detach_controller", 00:06:09.604 "bdev_nvme_get_controllers", 00:06:09.604 "bdev_nvme_attach_controller", 00:06:09.604 "bdev_nvme_set_hotplug", 00:06:09.604 "bdev_nvme_set_options", 00:06:09.604 "bdev_passthru_delete", 00:06:09.604 "bdev_passthru_create", 00:06:09.604 "bdev_lvol_set_parent_bdev", 00:06:09.604 "bdev_lvol_set_parent", 00:06:09.604 "bdev_lvol_check_shallow_copy", 00:06:09.604 "bdev_lvol_start_shallow_copy", 00:06:09.604 "bdev_lvol_grow_lvstore", 00:06:09.604 "bdev_lvol_get_lvols", 00:06:09.604 "bdev_lvol_get_lvstores", 00:06:09.604 "bdev_lvol_delete", 00:06:09.604 "bdev_lvol_set_read_only", 00:06:09.604 "bdev_lvol_resize", 00:06:09.604 "bdev_lvol_decouple_parent", 00:06:09.604 "bdev_lvol_inflate", 00:06:09.604 "bdev_lvol_rename", 00:06:09.604 "bdev_lvol_clone_bdev", 00:06:09.604 "bdev_lvol_clone", 00:06:09.604 "bdev_lvol_snapshot", 00:06:09.604 "bdev_lvol_create", 00:06:09.604 "bdev_lvol_delete_lvstore", 00:06:09.604 "bdev_lvol_rename_lvstore", 00:06:09.604 "bdev_lvol_create_lvstore", 00:06:09.604 "bdev_raid_set_options", 00:06:09.604 "bdev_raid_remove_base_bdev", 00:06:09.604 "bdev_raid_add_base_bdev", 00:06:09.604 "bdev_raid_delete", 00:06:09.604 "bdev_raid_create", 00:06:09.604 "bdev_raid_get_bdevs", 00:06:09.604 "bdev_error_inject_error", 00:06:09.604 "bdev_error_delete", 00:06:09.604 "bdev_error_create", 00:06:09.604 "bdev_split_delete", 00:06:09.604 "bdev_split_create", 00:06:09.604 "bdev_delay_delete", 00:06:09.604 "bdev_delay_create", 00:06:09.604 "bdev_delay_update_latency", 00:06:09.604 "bdev_zone_block_delete", 00:06:09.604 "bdev_zone_block_create", 00:06:09.604 "blobfs_create", 00:06:09.604 "blobfs_detect", 00:06:09.604 "blobfs_set_cache_size", 00:06:09.604 "bdev_aio_delete", 00:06:09.604 "bdev_aio_rescan", 00:06:09.604 "bdev_aio_create", 00:06:09.604 "bdev_ftl_set_property", 00:06:09.604 "bdev_ftl_get_properties", 00:06:09.604 "bdev_ftl_get_stats", 00:06:09.604 "bdev_ftl_unmap", 00:06:09.604 "bdev_ftl_unload", 00:06:09.604 "bdev_ftl_delete", 00:06:09.604 "bdev_ftl_load", 00:06:09.604 "bdev_ftl_create", 00:06:09.604 "bdev_virtio_attach_controller", 00:06:09.604 "bdev_virtio_scsi_get_devices", 00:06:09.604 "bdev_virtio_detach_controller", 00:06:09.604 "bdev_virtio_blk_set_hotplug", 00:06:09.604 "bdev_iscsi_delete", 00:06:09.604 "bdev_iscsi_create", 00:06:09.604 "bdev_iscsi_set_options", 00:06:09.604 "accel_error_inject_error", 00:06:09.604 "ioat_scan_accel_module", 00:06:09.604 "dsa_scan_accel_module", 00:06:09.604 "iaa_scan_accel_module", 00:06:09.604 "keyring_file_remove_key", 00:06:09.604 "keyring_file_add_key", 00:06:09.604 "keyring_linux_set_options", 00:06:09.604 "fsdev_aio_delete", 00:06:09.604 "fsdev_aio_create", 00:06:09.604 "iscsi_get_histogram", 00:06:09.604 "iscsi_enable_histogram", 00:06:09.604 "iscsi_set_options", 00:06:09.604 "iscsi_get_auth_groups", 00:06:09.604 "iscsi_auth_group_remove_secret", 00:06:09.604 "iscsi_auth_group_add_secret", 00:06:09.604 "iscsi_delete_auth_group", 00:06:09.604 "iscsi_create_auth_group", 00:06:09.604 "iscsi_set_discovery_auth", 00:06:09.604 "iscsi_get_options", 00:06:09.604 "iscsi_target_node_request_logout", 00:06:09.604 "iscsi_target_node_set_redirect", 00:06:09.604 "iscsi_target_node_set_auth", 00:06:09.604 "iscsi_target_node_add_lun", 00:06:09.604 "iscsi_get_stats", 00:06:09.604 "iscsi_get_connections", 00:06:09.604 "iscsi_portal_group_set_auth", 00:06:09.604 "iscsi_start_portal_group", 00:06:09.604 "iscsi_delete_portal_group", 00:06:09.604 "iscsi_create_portal_group", 00:06:09.604 "iscsi_get_portal_groups", 00:06:09.604 "iscsi_delete_target_node", 00:06:09.604 "iscsi_target_node_remove_pg_ig_maps", 00:06:09.604 "iscsi_target_node_add_pg_ig_maps", 00:06:09.604 "iscsi_create_target_node", 00:06:09.604 "iscsi_get_target_nodes", 00:06:09.604 "iscsi_delete_initiator_group", 00:06:09.604 "iscsi_initiator_group_remove_initiators", 00:06:09.604 "iscsi_initiator_group_add_initiators", 00:06:09.604 "iscsi_create_initiator_group", 00:06:09.604 "iscsi_get_initiator_groups", 00:06:09.604 "nvmf_set_crdt", 00:06:09.604 "nvmf_set_config", 00:06:09.604 "nvmf_set_max_subsystems", 00:06:09.604 "nvmf_stop_mdns_prr", 00:06:09.604 "nvmf_publish_mdns_prr", 00:06:09.604 "nvmf_subsystem_get_listeners", 00:06:09.604 "nvmf_subsystem_get_qpairs", 00:06:09.604 "nvmf_subsystem_get_controllers", 00:06:09.604 "nvmf_get_stats", 00:06:09.604 "nvmf_get_transports", 00:06:09.604 "nvmf_create_transport", 00:06:09.604 "nvmf_get_targets", 00:06:09.604 "nvmf_delete_target", 00:06:09.604 "nvmf_create_target", 00:06:09.604 "nvmf_subsystem_allow_any_host", 00:06:09.604 "nvmf_subsystem_set_keys", 00:06:09.604 "nvmf_subsystem_remove_host", 00:06:09.604 "nvmf_subsystem_add_host", 00:06:09.604 "nvmf_ns_remove_host", 00:06:09.604 "nvmf_ns_add_host", 00:06:09.604 "nvmf_subsystem_remove_ns", 00:06:09.604 "nvmf_subsystem_set_ns_ana_group", 00:06:09.604 "nvmf_subsystem_add_ns", 00:06:09.604 "nvmf_subsystem_listener_set_ana_state", 00:06:09.604 "nvmf_discovery_get_referrals", 00:06:09.604 "nvmf_discovery_remove_referral", 00:06:09.604 "nvmf_discovery_add_referral", 00:06:09.604 "nvmf_subsystem_remove_listener", 00:06:09.604 "nvmf_subsystem_add_listener", 00:06:09.604 "nvmf_delete_subsystem", 00:06:09.604 "nvmf_create_subsystem", 00:06:09.604 "nvmf_get_subsystems", 00:06:09.604 "env_dpdk_get_mem_stats", 00:06:09.604 "nbd_get_disks", 00:06:09.604 "nbd_stop_disk", 00:06:09.604 "nbd_start_disk", 00:06:09.604 "ublk_recover_disk", 00:06:09.604 "ublk_get_disks", 00:06:09.604 "ublk_stop_disk", 00:06:09.604 "ublk_start_disk", 00:06:09.604 "ublk_destroy_target", 00:06:09.604 "ublk_create_target", 00:06:09.604 "virtio_blk_create_transport", 00:06:09.604 "virtio_blk_get_transports", 00:06:09.604 "vhost_controller_set_coalescing", 00:06:09.604 "vhost_get_controllers", 00:06:09.604 "vhost_delete_controller", 00:06:09.604 "vhost_create_blk_controller", 00:06:09.604 "vhost_scsi_controller_remove_target", 00:06:09.604 "vhost_scsi_controller_add_target", 00:06:09.604 "vhost_start_scsi_controller", 00:06:09.604 "vhost_create_scsi_controller", 00:06:09.604 "thread_set_cpumask", 00:06:09.604 "scheduler_set_options", 00:06:09.604 "framework_get_governor", 00:06:09.604 "framework_get_scheduler", 00:06:09.604 "framework_set_scheduler", 00:06:09.604 "framework_get_reactors", 00:06:09.604 "thread_get_io_channels", 00:06:09.604 "thread_get_pollers", 00:06:09.604 "thread_get_stats", 00:06:09.604 "framework_monitor_context_switch", 00:06:09.604 "spdk_kill_instance", 00:06:09.604 "log_enable_timestamps", 00:06:09.604 "log_get_flags", 00:06:09.604 "log_clear_flag", 00:06:09.604 "log_set_flag", 00:06:09.604 "log_get_level", 00:06:09.604 "log_set_level", 00:06:09.604 "log_get_print_level", 00:06:09.604 "log_set_print_level", 00:06:09.604 "framework_enable_cpumask_locks", 00:06:09.604 "framework_disable_cpumask_locks", 00:06:09.604 "framework_wait_init", 00:06:09.604 "framework_start_init", 00:06:09.604 "scsi_get_devices", 00:06:09.604 "bdev_get_histogram", 00:06:09.604 "bdev_enable_histogram", 00:06:09.604 "bdev_set_qos_limit", 00:06:09.604 "bdev_set_qd_sampling_period", 00:06:09.604 "bdev_get_bdevs", 00:06:09.604 "bdev_reset_iostat", 00:06:09.604 "bdev_get_iostat", 00:06:09.604 "bdev_examine", 00:06:09.604 "bdev_wait_for_examine", 00:06:09.604 "bdev_set_options", 00:06:09.604 "accel_get_stats", 00:06:09.604 "accel_set_options", 00:06:09.605 "accel_set_driver", 00:06:09.605 "accel_crypto_key_destroy", 00:06:09.605 "accel_crypto_keys_get", 00:06:09.605 "accel_crypto_key_create", 00:06:09.605 "accel_assign_opc", 00:06:09.605 "accel_get_module_info", 00:06:09.605 "accel_get_opc_assignments", 00:06:09.605 "vmd_rescan", 00:06:09.605 "vmd_remove_device", 00:06:09.605 "vmd_enable", 00:06:09.605 "sock_get_default_impl", 00:06:09.605 "sock_set_default_impl", 00:06:09.605 "sock_impl_set_options", 00:06:09.605 "sock_impl_get_options", 00:06:09.605 "iobuf_get_stats", 00:06:09.605 "iobuf_set_options", 00:06:09.605 "keyring_get_keys", 00:06:09.605 "framework_get_pci_devices", 00:06:09.605 "framework_get_config", 00:06:09.605 "framework_get_subsystems", 00:06:09.605 "fsdev_set_opts", 00:06:09.605 "fsdev_get_opts", 00:06:09.605 "trace_get_info", 00:06:09.605 "trace_get_tpoint_group_mask", 00:06:09.605 "trace_disable_tpoint_group", 00:06:09.605 "trace_enable_tpoint_group", 00:06:09.605 "trace_clear_tpoint_mask", 00:06:09.605 "trace_set_tpoint_mask", 00:06:09.605 "notify_get_notifications", 00:06:09.605 "notify_get_types", 00:06:09.605 "spdk_get_version", 00:06:09.605 "rpc_get_methods" 00:06:09.605 ] 00:06:09.605 18:56:27 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:06:09.605 18:56:27 spdkcli_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:06:09.605 18:56:27 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:09.605 18:56:27 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:06:09.605 18:56:27 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 69619 00:06:09.605 18:56:27 spdkcli_tcp -- common/autotest_common.sh@954 -- # '[' -z 69619 ']' 00:06:09.605 18:56:27 spdkcli_tcp -- common/autotest_common.sh@958 -- # kill -0 69619 00:06:09.605 18:56:27 spdkcli_tcp -- common/autotest_common.sh@959 -- # uname 00:06:09.605 18:56:27 spdkcli_tcp -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:09.605 18:56:27 spdkcli_tcp -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69619 00:06:09.605 killing process with pid 69619 00:06:09.605 18:56:27 spdkcli_tcp -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:09.605 18:56:27 spdkcli_tcp -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:09.605 18:56:27 spdkcli_tcp -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69619' 00:06:09.605 18:56:27 spdkcli_tcp -- common/autotest_common.sh@973 -- # kill 69619 00:06:09.605 18:56:27 spdkcli_tcp -- common/autotest_common.sh@978 -- # wait 69619 00:06:10.175 ************************************ 00:06:10.175 END TEST spdkcli_tcp 00:06:10.175 ************************************ 00:06:10.175 00:06:10.175 real 0m1.787s 00:06:10.175 user 0m2.969s 00:06:10.175 sys 0m0.572s 00:06:10.175 18:56:27 spdkcli_tcp -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:10.175 18:56:27 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:10.175 18:56:27 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:06:10.175 18:56:27 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:10.175 18:56:27 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:10.175 18:56:27 -- common/autotest_common.sh@10 -- # set +x 00:06:10.175 ************************************ 00:06:10.175 START TEST dpdk_mem_utility 00:06:10.175 ************************************ 00:06:10.175 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:06:10.175 * Looking for test storage... 00:06:10.175 * Found test storage at /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility 00:06:10.175 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:06:10.175 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@1711 -- # lcov --version 00:06:10.175 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:06:10.435 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:10.435 18:56:27 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:06:10.435 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:10.435 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:06:10.435 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.435 --rc genhtml_branch_coverage=1 00:06:10.435 --rc genhtml_function_coverage=1 00:06:10.435 --rc genhtml_legend=1 00:06:10.435 --rc geninfo_all_blocks=1 00:06:10.435 --rc geninfo_unexecuted_blocks=1 00:06:10.435 00:06:10.435 ' 00:06:10.435 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:06:10.435 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.435 --rc genhtml_branch_coverage=1 00:06:10.435 --rc genhtml_function_coverage=1 00:06:10.435 --rc genhtml_legend=1 00:06:10.435 --rc geninfo_all_blocks=1 00:06:10.435 --rc geninfo_unexecuted_blocks=1 00:06:10.435 00:06:10.435 ' 00:06:10.435 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:06:10.435 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.435 --rc genhtml_branch_coverage=1 00:06:10.435 --rc genhtml_function_coverage=1 00:06:10.435 --rc genhtml_legend=1 00:06:10.435 --rc geninfo_all_blocks=1 00:06:10.435 --rc geninfo_unexecuted_blocks=1 00:06:10.435 00:06:10.435 ' 00:06:10.435 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:06:10.435 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.435 --rc genhtml_branch_coverage=1 00:06:10.436 --rc genhtml_function_coverage=1 00:06:10.436 --rc genhtml_legend=1 00:06:10.436 --rc geninfo_all_blocks=1 00:06:10.436 --rc geninfo_unexecuted_blocks=1 00:06:10.436 00:06:10.436 ' 00:06:10.436 18:56:27 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:06:10.436 18:56:27 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=69719 00:06:10.436 18:56:27 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:10.436 18:56:27 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 69719 00:06:10.436 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@835 -- # '[' -z 69719 ']' 00:06:10.436 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:10.436 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:10.436 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:10.436 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:10.436 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:10.436 18:56:27 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:06:10.436 [2024-12-05 18:56:27.861040] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:10.436 [2024-12-05 18:56:27.861680] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69719 ] 00:06:10.696 [2024-12-05 18:56:28.019261] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:10.696 [2024-12-05 18:56:28.043583] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:11.269 18:56:28 dpdk_mem_utility -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:11.269 18:56:28 dpdk_mem_utility -- common/autotest_common.sh@868 -- # return 0 00:06:11.269 18:56:28 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:06:11.269 18:56:28 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:06:11.270 18:56:28 dpdk_mem_utility -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:11.270 18:56:28 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:06:11.270 { 00:06:11.270 "filename": "/tmp/spdk_mem_dump.txt" 00:06:11.270 } 00:06:11.270 18:56:28 dpdk_mem_utility -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:11.270 18:56:28 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:06:11.270 DPDK memory size 818.000000 MiB in 1 heap(s) 00:06:11.270 1 heaps totaling size 818.000000 MiB 00:06:11.270 size: 818.000000 MiB heap id: 0 00:06:11.270 end heaps---------- 00:06:11.270 9 mempools totaling size 603.782043 MiB 00:06:11.270 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:06:11.270 size: 158.602051 MiB name: PDU_data_out_Pool 00:06:11.270 size: 100.555481 MiB name: bdev_io_69719 00:06:11.270 size: 50.003479 MiB name: msgpool_69719 00:06:11.270 size: 36.509338 MiB name: fsdev_io_69719 00:06:11.270 size: 21.763794 MiB name: PDU_Pool 00:06:11.270 size: 19.513306 MiB name: SCSI_TASK_Pool 00:06:11.270 size: 4.133484 MiB name: evtpool_69719 00:06:11.270 size: 0.026123 MiB name: Session_Pool 00:06:11.270 end mempools------- 00:06:11.270 6 memzones totaling size 4.142822 MiB 00:06:11.270 size: 1.000366 MiB name: RG_ring_0_69719 00:06:11.270 size: 1.000366 MiB name: RG_ring_1_69719 00:06:11.270 size: 1.000366 MiB name: RG_ring_4_69719 00:06:11.270 size: 1.000366 MiB name: RG_ring_5_69719 00:06:11.270 size: 0.125366 MiB name: RG_ring_2_69719 00:06:11.270 size: 0.015991 MiB name: RG_ring_3_69719 00:06:11.270 end memzones------- 00:06:11.270 18:56:28 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py -m 0 00:06:11.270 heap id: 0 total size: 818.000000 MiB number of busy elements: 309 number of free elements: 15 00:06:11.270 list of free elements. size: 10.803955 MiB 00:06:11.270 element at address: 0x200019200000 with size: 0.999878 MiB 00:06:11.270 element at address: 0x200019400000 with size: 0.999878 MiB 00:06:11.270 element at address: 0x200032000000 with size: 0.994446 MiB 00:06:11.270 element at address: 0x200000400000 with size: 0.993958 MiB 00:06:11.270 element at address: 0x200006400000 with size: 0.959839 MiB 00:06:11.270 element at address: 0x200012c00000 with size: 0.944275 MiB 00:06:11.270 element at address: 0x200019600000 with size: 0.936584 MiB 00:06:11.270 element at address: 0x200000200000 with size: 0.717346 MiB 00:06:11.270 element at address: 0x20001ae00000 with size: 0.568420 MiB 00:06:11.270 element at address: 0x20000a600000 with size: 0.488892 MiB 00:06:11.270 element at address: 0x200000c00000 with size: 0.486267 MiB 00:06:11.270 element at address: 0x200019800000 with size: 0.485657 MiB 00:06:11.270 element at address: 0x200003e00000 with size: 0.480286 MiB 00:06:11.270 element at address: 0x200028200000 with size: 0.396484 MiB 00:06:11.270 element at address: 0x200000800000 with size: 0.351746 MiB 00:06:11.270 list of standard malloc elements. size: 199.267151 MiB 00:06:11.270 element at address: 0x20000a7fff80 with size: 132.000122 MiB 00:06:11.270 element at address: 0x2000065fff80 with size: 64.000122 MiB 00:06:11.270 element at address: 0x2000192fff80 with size: 1.000122 MiB 00:06:11.270 element at address: 0x2000194fff80 with size: 1.000122 MiB 00:06:11.270 element at address: 0x2000196fff80 with size: 1.000122 MiB 00:06:11.270 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:06:11.270 element at address: 0x2000196eff00 with size: 0.062622 MiB 00:06:11.270 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:06:11.270 element at address: 0x2000196efdc0 with size: 0.000305 MiB 00:06:11.270 element at address: 0x2000002d7c40 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000003d9e40 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004fe740 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004fe800 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004fe8c0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004fe980 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004fea40 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004feb00 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004febc0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004fec80 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004fed40 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004fee00 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004feec0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004fef80 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ff040 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ff100 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ff1c0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ff280 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ff340 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ff400 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ff4c0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ff580 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ff640 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ff700 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ff7c0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ff880 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ff940 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ffa00 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ffac0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ffcc0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ffd80 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000004ffe40 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000085a0c0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000085a2c0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000085e580 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087e840 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087e900 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087e9c0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087ea80 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087eb40 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087ec00 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087ecc0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087ed80 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087ee40 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087ef00 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087efc0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087f080 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087f140 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087f200 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087f2c0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087f380 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087f440 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087f500 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087f5c0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x20000087f680 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000008ff940 with size: 0.000183 MiB 00:06:11.270 element at address: 0x2000008ffb40 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7c7c0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7c880 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7c940 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7ca00 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7cac0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7cb80 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7cc40 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7cd00 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7cdc0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7ce80 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7cf40 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7d000 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7d0c0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7d180 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7d240 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7d300 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7d3c0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7d480 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7d540 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7d600 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7d6c0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7d780 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7d840 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7d900 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7d9c0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7da80 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7db40 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7dc00 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7dcc0 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7dd80 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7de40 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7df00 with size: 0.000183 MiB 00:06:11.270 element at address: 0x200000c7dfc0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7e080 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7e140 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7e200 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7e2c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7e380 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7e440 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7e500 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7e5c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7e680 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7e740 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7e800 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7e8c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7e980 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7ea40 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7eb00 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7ebc0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7ec80 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000c7ed40 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000cff000 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200000cff0c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200003e7af40 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200003e7b000 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200003e7b0c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200003e7b180 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200003e7b240 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200003e7b300 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200003e7b3c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200003e7b480 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200003e7b540 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200003e7b600 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200003e7b6c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200003efb980 with size: 0.000183 MiB 00:06:11.271 element at address: 0x2000064fdd80 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20000a67d280 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20000a67d340 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20000a67d400 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20000a67d4c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20000a67d580 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20000a67d640 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20000a67d700 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20000a67d7c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20000a67d880 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20000a67d940 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20000a67da00 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20000a67dac0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20000a6fdd80 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200012cf1bc0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x2000196efc40 with size: 0.000183 MiB 00:06:11.271 element at address: 0x2000196efd00 with size: 0.000183 MiB 00:06:11.271 element at address: 0x2000198bc740 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae91840 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae91900 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae919c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae91a80 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae91b40 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae91c00 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae91cc0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae91d80 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae91e40 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae91f00 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae91fc0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92080 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92140 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92200 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae922c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92380 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92440 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92500 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae925c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92680 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92740 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92800 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae928c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92980 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92a40 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92b00 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92bc0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92c80 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92d40 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92e00 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92ec0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae92f80 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93040 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93100 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae931c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93280 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93340 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93400 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae934c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93580 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93640 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93700 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae937c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93880 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93940 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93a00 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93ac0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93b80 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93c40 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93d00 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93dc0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93e80 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae93f40 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94000 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae940c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94180 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94240 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94300 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae943c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94480 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94540 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94600 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae946c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94780 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94840 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94900 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae949c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94a80 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94b40 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94c00 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94cc0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94d80 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94e40 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94f00 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae94fc0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae95080 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae95140 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae95200 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae952c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae95380 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20001ae95440 with size: 0.000183 MiB 00:06:11.271 element at address: 0x200028265800 with size: 0.000183 MiB 00:06:11.271 element at address: 0x2000282658c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20002826c4c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20002826c6c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20002826c780 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20002826c840 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20002826c900 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20002826c9c0 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20002826ca80 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20002826cb40 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20002826cc00 with size: 0.000183 MiB 00:06:11.271 element at address: 0x20002826ccc0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826cd80 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826ce40 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826cf00 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826cfc0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826d080 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826d140 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826d200 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826d2c0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826d380 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826d440 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826d500 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826d5c0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826d680 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826d740 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826d800 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826d8c0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826d980 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826da40 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826db00 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826dbc0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826dc80 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826dd40 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826de00 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826dec0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826df80 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826e040 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826e100 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826e1c0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826e280 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826e340 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826e400 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826e4c0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826e580 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826e640 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826e700 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826e7c0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826e880 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826e940 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826ea00 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826eac0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826eb80 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826ec40 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826ed00 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826edc0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826ee80 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826ef40 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826f000 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826f0c0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826f180 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826f240 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826f300 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826f3c0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826f480 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826f540 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826f600 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826f6c0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826f780 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826f840 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826f900 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826f9c0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826fa80 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826fb40 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826fc00 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826fcc0 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826fd80 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826fe40 with size: 0.000183 MiB 00:06:11.272 element at address: 0x20002826ff00 with size: 0.000183 MiB 00:06:11.272 list of memzone associated elements. size: 607.928894 MiB 00:06:11.272 element at address: 0x20001ae95500 with size: 211.416748 MiB 00:06:11.272 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:06:11.272 element at address: 0x20002826ffc0 with size: 157.562561 MiB 00:06:11.272 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:06:11.272 element at address: 0x200012df1e80 with size: 100.055054 MiB 00:06:11.272 associated memzone info: size: 100.054932 MiB name: MP_bdev_io_69719_0 00:06:11.272 element at address: 0x200000dff380 with size: 48.003052 MiB 00:06:11.272 associated memzone info: size: 48.002930 MiB name: MP_msgpool_69719_0 00:06:11.272 element at address: 0x200003ffdb80 with size: 36.008911 MiB 00:06:11.272 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_69719_0 00:06:11.272 element at address: 0x2000199be940 with size: 20.255554 MiB 00:06:11.272 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:06:11.272 element at address: 0x2000321feb40 with size: 18.005066 MiB 00:06:11.272 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:06:11.272 element at address: 0x2000004fff00 with size: 3.000244 MiB 00:06:11.272 associated memzone info: size: 3.000122 MiB name: MP_evtpool_69719_0 00:06:11.272 element at address: 0x2000009ffe00 with size: 2.000488 MiB 00:06:11.272 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_69719 00:06:11.272 element at address: 0x2000002d7d00 with size: 1.008118 MiB 00:06:11.272 associated memzone info: size: 1.007996 MiB name: MP_evtpool_69719 00:06:11.272 element at address: 0x20000a6fde40 with size: 1.008118 MiB 00:06:11.272 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:06:11.272 element at address: 0x2000198bc800 with size: 1.008118 MiB 00:06:11.272 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:06:11.272 element at address: 0x2000064fde40 with size: 1.008118 MiB 00:06:11.272 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:06:11.272 element at address: 0x200003efba40 with size: 1.008118 MiB 00:06:11.272 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:06:11.272 element at address: 0x200000cff180 with size: 1.000488 MiB 00:06:11.272 associated memzone info: size: 1.000366 MiB name: RG_ring_0_69719 00:06:11.272 element at address: 0x2000008ffc00 with size: 1.000488 MiB 00:06:11.272 associated memzone info: size: 1.000366 MiB name: RG_ring_1_69719 00:06:11.272 element at address: 0x200012cf1c80 with size: 1.000488 MiB 00:06:11.272 associated memzone info: size: 1.000366 MiB name: RG_ring_4_69719 00:06:11.272 element at address: 0x2000320fe940 with size: 1.000488 MiB 00:06:11.272 associated memzone info: size: 1.000366 MiB name: RG_ring_5_69719 00:06:11.272 element at address: 0x20000087f740 with size: 0.500488 MiB 00:06:11.272 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_69719 00:06:11.272 element at address: 0x200000c7ee00 with size: 0.500488 MiB 00:06:11.272 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_69719 00:06:11.272 element at address: 0x20000a67db80 with size: 0.500488 MiB 00:06:11.272 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:06:11.272 element at address: 0x200003e7b780 with size: 0.500488 MiB 00:06:11.272 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:06:11.272 element at address: 0x20001987c540 with size: 0.250488 MiB 00:06:11.272 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:06:11.272 element at address: 0x2000002b7a40 with size: 0.125488 MiB 00:06:11.272 associated memzone info: size: 0.125366 MiB name: RG_MP_evtpool_69719 00:06:11.272 element at address: 0x20000085e640 with size: 0.125488 MiB 00:06:11.272 associated memzone info: size: 0.125366 MiB name: RG_ring_2_69719 00:06:11.272 element at address: 0x2000064f5b80 with size: 0.031738 MiB 00:06:11.272 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:06:11.272 element at address: 0x200028265980 with size: 0.023743 MiB 00:06:11.272 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:06:11.272 element at address: 0x20000085a380 with size: 0.016113 MiB 00:06:11.272 associated memzone info: size: 0.015991 MiB name: RG_ring_3_69719 00:06:11.272 element at address: 0x20002826bac0 with size: 0.002441 MiB 00:06:11.272 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:06:11.272 element at address: 0x2000004ffb80 with size: 0.000305 MiB 00:06:11.272 associated memzone info: size: 0.000183 MiB name: MP_msgpool_69719 00:06:11.272 element at address: 0x2000008ffa00 with size: 0.000305 MiB 00:06:11.272 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_69719 00:06:11.272 element at address: 0x20000085a180 with size: 0.000305 MiB 00:06:11.272 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_69719 00:06:11.272 element at address: 0x20002826c580 with size: 0.000305 MiB 00:06:11.272 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:06:11.272 18:56:28 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:06:11.272 18:56:28 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 69719 00:06:11.272 18:56:28 dpdk_mem_utility -- common/autotest_common.sh@954 -- # '[' -z 69719 ']' 00:06:11.272 18:56:28 dpdk_mem_utility -- common/autotest_common.sh@958 -- # kill -0 69719 00:06:11.272 18:56:28 dpdk_mem_utility -- common/autotest_common.sh@959 -- # uname 00:06:11.273 18:56:28 dpdk_mem_utility -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:11.273 18:56:28 dpdk_mem_utility -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69719 00:06:11.273 18:56:28 dpdk_mem_utility -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:11.273 18:56:28 dpdk_mem_utility -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:11.273 18:56:28 dpdk_mem_utility -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69719' 00:06:11.273 killing process with pid 69719 00:06:11.273 18:56:28 dpdk_mem_utility -- common/autotest_common.sh@973 -- # kill 69719 00:06:11.273 18:56:28 dpdk_mem_utility -- common/autotest_common.sh@978 -- # wait 69719 00:06:11.844 00:06:11.844 real 0m1.622s 00:06:11.844 user 0m1.535s 00:06:11.844 sys 0m0.510s 00:06:11.844 18:56:29 dpdk_mem_utility -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:11.844 18:56:29 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:06:11.844 ************************************ 00:06:11.844 END TEST dpdk_mem_utility 00:06:11.844 ************************************ 00:06:11.844 18:56:29 -- spdk/autotest.sh@168 -- # run_test event /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:06:11.844 18:56:29 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:11.844 18:56:29 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:11.844 18:56:29 -- common/autotest_common.sh@10 -- # set +x 00:06:11.844 ************************************ 00:06:11.844 START TEST event 00:06:11.844 ************************************ 00:06:11.844 18:56:29 event -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:06:11.845 * Looking for test storage... 00:06:11.845 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:06:11.845 18:56:29 event -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:06:11.845 18:56:29 event -- common/autotest_common.sh@1711 -- # lcov --version 00:06:11.845 18:56:29 event -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:06:12.105 18:56:29 event -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:06:12.105 18:56:29 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:12.105 18:56:29 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:12.105 18:56:29 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:12.105 18:56:29 event -- scripts/common.sh@336 -- # IFS=.-: 00:06:12.105 18:56:29 event -- scripts/common.sh@336 -- # read -ra ver1 00:06:12.105 18:56:29 event -- scripts/common.sh@337 -- # IFS=.-: 00:06:12.105 18:56:29 event -- scripts/common.sh@337 -- # read -ra ver2 00:06:12.105 18:56:29 event -- scripts/common.sh@338 -- # local 'op=<' 00:06:12.105 18:56:29 event -- scripts/common.sh@340 -- # ver1_l=2 00:06:12.105 18:56:29 event -- scripts/common.sh@341 -- # ver2_l=1 00:06:12.105 18:56:29 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:12.105 18:56:29 event -- scripts/common.sh@344 -- # case "$op" in 00:06:12.105 18:56:29 event -- scripts/common.sh@345 -- # : 1 00:06:12.105 18:56:29 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:12.105 18:56:29 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:12.105 18:56:29 event -- scripts/common.sh@365 -- # decimal 1 00:06:12.105 18:56:29 event -- scripts/common.sh@353 -- # local d=1 00:06:12.105 18:56:29 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:12.105 18:56:29 event -- scripts/common.sh@355 -- # echo 1 00:06:12.105 18:56:29 event -- scripts/common.sh@365 -- # ver1[v]=1 00:06:12.105 18:56:29 event -- scripts/common.sh@366 -- # decimal 2 00:06:12.105 18:56:29 event -- scripts/common.sh@353 -- # local d=2 00:06:12.105 18:56:29 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:12.105 18:56:29 event -- scripts/common.sh@355 -- # echo 2 00:06:12.105 18:56:29 event -- scripts/common.sh@366 -- # ver2[v]=2 00:06:12.105 18:56:29 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:12.105 18:56:29 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:12.105 18:56:29 event -- scripts/common.sh@368 -- # return 0 00:06:12.105 18:56:29 event -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:12.105 18:56:29 event -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:06:12.105 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:12.105 --rc genhtml_branch_coverage=1 00:06:12.105 --rc genhtml_function_coverage=1 00:06:12.105 --rc genhtml_legend=1 00:06:12.105 --rc geninfo_all_blocks=1 00:06:12.105 --rc geninfo_unexecuted_blocks=1 00:06:12.105 00:06:12.105 ' 00:06:12.105 18:56:29 event -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:06:12.105 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:12.105 --rc genhtml_branch_coverage=1 00:06:12.105 --rc genhtml_function_coverage=1 00:06:12.105 --rc genhtml_legend=1 00:06:12.105 --rc geninfo_all_blocks=1 00:06:12.105 --rc geninfo_unexecuted_blocks=1 00:06:12.105 00:06:12.105 ' 00:06:12.105 18:56:29 event -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:06:12.106 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:12.106 --rc genhtml_branch_coverage=1 00:06:12.106 --rc genhtml_function_coverage=1 00:06:12.106 --rc genhtml_legend=1 00:06:12.106 --rc geninfo_all_blocks=1 00:06:12.106 --rc geninfo_unexecuted_blocks=1 00:06:12.106 00:06:12.106 ' 00:06:12.106 18:56:29 event -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:06:12.106 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:12.106 --rc genhtml_branch_coverage=1 00:06:12.106 --rc genhtml_function_coverage=1 00:06:12.106 --rc genhtml_legend=1 00:06:12.106 --rc geninfo_all_blocks=1 00:06:12.106 --rc geninfo_unexecuted_blocks=1 00:06:12.106 00:06:12.106 ' 00:06:12.106 18:56:29 event -- event/event.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:06:12.106 18:56:29 event -- bdev/nbd_common.sh@6 -- # set -e 00:06:12.106 18:56:29 event -- event/event.sh@45 -- # run_test event_perf /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:06:12.106 18:56:29 event -- common/autotest_common.sh@1105 -- # '[' 6 -le 1 ']' 00:06:12.106 18:56:29 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:12.106 18:56:29 event -- common/autotest_common.sh@10 -- # set +x 00:06:12.106 ************************************ 00:06:12.106 START TEST event_perf 00:06:12.106 ************************************ 00:06:12.106 18:56:29 event.event_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:06:12.106 Running I/O for 1 seconds...[2024-12-05 18:56:29.512641] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:12.106 [2024-12-05 18:56:29.512771] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69805 ] 00:06:12.366 [2024-12-05 18:56:29.667455] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:06:12.366 [2024-12-05 18:56:29.695002] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:12.366 [2024-12-05 18:56:29.695234] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:12.366 [2024-12-05 18:56:29.695296] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:12.366 Running I/O for 1 seconds...[2024-12-05 18:56:29.695443] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:06:13.305 00:06:13.305 lcore 0: 215172 00:06:13.305 lcore 1: 215173 00:06:13.305 lcore 2: 215173 00:06:13.305 lcore 3: 215174 00:06:13.305 done. 00:06:13.305 00:06:13.305 real 0m1.291s 00:06:13.305 user 0m4.062s 00:06:13.305 sys 0m0.112s 00:06:13.305 18:56:30 event.event_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:13.305 18:56:30 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:06:13.305 ************************************ 00:06:13.305 END TEST event_perf 00:06:13.305 ************************************ 00:06:13.305 18:56:30 event -- event/event.sh@46 -- # run_test event_reactor /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:06:13.305 18:56:30 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:06:13.305 18:56:30 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:13.305 18:56:30 event -- common/autotest_common.sh@10 -- # set +x 00:06:13.305 ************************************ 00:06:13.305 START TEST event_reactor 00:06:13.305 ************************************ 00:06:13.305 18:56:30 event.event_reactor -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:06:13.566 [2024-12-05 18:56:30.871148] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:13.566 [2024-12-05 18:56:30.871287] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69839 ] 00:06:13.566 [2024-12-05 18:56:31.026211] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:13.566 [2024-12-05 18:56:31.050347] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:14.948 test_start 00:06:14.948 oneshot 00:06:14.948 tick 100 00:06:14.948 tick 100 00:06:14.948 tick 250 00:06:14.948 tick 100 00:06:14.948 tick 100 00:06:14.948 tick 100 00:06:14.948 tick 250 00:06:14.948 tick 500 00:06:14.948 tick 100 00:06:14.948 tick 100 00:06:14.948 tick 250 00:06:14.948 tick 100 00:06:14.948 tick 100 00:06:14.948 test_end 00:06:14.948 00:06:14.948 real 0m1.277s 00:06:14.948 user 0m1.093s 00:06:14.948 sys 0m0.078s 00:06:14.948 18:56:32 event.event_reactor -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:14.948 18:56:32 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:06:14.948 ************************************ 00:06:14.948 END TEST event_reactor 00:06:14.948 ************************************ 00:06:14.948 18:56:32 event -- event/event.sh@47 -- # run_test event_reactor_perf /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:06:14.948 18:56:32 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:06:14.948 18:56:32 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:14.948 18:56:32 event -- common/autotest_common.sh@10 -- # set +x 00:06:14.948 ************************************ 00:06:14.948 START TEST event_reactor_perf 00:06:14.948 ************************************ 00:06:14.948 18:56:32 event.event_reactor_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:06:14.948 [2024-12-05 18:56:32.214335] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:14.948 [2024-12-05 18:56:32.214495] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69870 ] 00:06:14.948 [2024-12-05 18:56:32.371216] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:14.948 [2024-12-05 18:56:32.395697] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:15.890 test_start 00:06:15.890 test_end 00:06:15.890 Performance: 413736 events per second 00:06:16.149 00:06:16.149 real 0m1.280s 00:06:16.149 user 0m1.102s 00:06:16.149 sys 0m0.072s 00:06:16.149 18:56:33 event.event_reactor_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:16.149 18:56:33 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:06:16.149 ************************************ 00:06:16.149 END TEST event_reactor_perf 00:06:16.149 ************************************ 00:06:16.149 18:56:33 event -- event/event.sh@49 -- # uname -s 00:06:16.149 18:56:33 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:06:16.149 18:56:33 event -- event/event.sh@50 -- # run_test event_scheduler /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:06:16.149 18:56:33 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:16.149 18:56:33 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:16.149 18:56:33 event -- common/autotest_common.sh@10 -- # set +x 00:06:16.149 ************************************ 00:06:16.149 START TEST event_scheduler 00:06:16.149 ************************************ 00:06:16.149 18:56:33 event.event_scheduler -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:06:16.149 * Looking for test storage... 00:06:16.149 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event/scheduler 00:06:16.149 18:56:33 event.event_scheduler -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:06:16.149 18:56:33 event.event_scheduler -- common/autotest_common.sh@1711 -- # lcov --version 00:06:16.149 18:56:33 event.event_scheduler -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:06:16.408 18:56:33 event.event_scheduler -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:16.408 18:56:33 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:06:16.408 18:56:33 event.event_scheduler -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:16.408 18:56:33 event.event_scheduler -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:06:16.408 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:16.408 --rc genhtml_branch_coverage=1 00:06:16.408 --rc genhtml_function_coverage=1 00:06:16.408 --rc genhtml_legend=1 00:06:16.408 --rc geninfo_all_blocks=1 00:06:16.408 --rc geninfo_unexecuted_blocks=1 00:06:16.408 00:06:16.408 ' 00:06:16.408 18:56:33 event.event_scheduler -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:06:16.408 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:16.408 --rc genhtml_branch_coverage=1 00:06:16.408 --rc genhtml_function_coverage=1 00:06:16.408 --rc genhtml_legend=1 00:06:16.408 --rc geninfo_all_blocks=1 00:06:16.408 --rc geninfo_unexecuted_blocks=1 00:06:16.408 00:06:16.408 ' 00:06:16.408 18:56:33 event.event_scheduler -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:06:16.408 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:16.408 --rc genhtml_branch_coverage=1 00:06:16.408 --rc genhtml_function_coverage=1 00:06:16.408 --rc genhtml_legend=1 00:06:16.408 --rc geninfo_all_blocks=1 00:06:16.408 --rc geninfo_unexecuted_blocks=1 00:06:16.408 00:06:16.408 ' 00:06:16.408 18:56:33 event.event_scheduler -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:06:16.408 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:16.408 --rc genhtml_branch_coverage=1 00:06:16.408 --rc genhtml_function_coverage=1 00:06:16.408 --rc genhtml_legend=1 00:06:16.408 --rc geninfo_all_blocks=1 00:06:16.408 --rc geninfo_unexecuted_blocks=1 00:06:16.408 00:06:16.408 ' 00:06:16.408 18:56:33 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:06:16.408 18:56:33 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:06:16.408 18:56:33 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=69946 00:06:16.408 18:56:33 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:06:16.408 18:56:33 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 69946 00:06:16.408 18:56:33 event.event_scheduler -- common/autotest_common.sh@835 -- # '[' -z 69946 ']' 00:06:16.408 18:56:33 event.event_scheduler -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:16.408 18:56:33 event.event_scheduler -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:16.408 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:16.408 18:56:33 event.event_scheduler -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:16.408 18:56:33 event.event_scheduler -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:16.408 18:56:33 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:16.408 [2024-12-05 18:56:33.829996] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:16.409 [2024-12-05 18:56:33.830115] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69946 ] 00:06:16.668 [2024-12-05 18:56:33.988177] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:06:16.668 [2024-12-05 18:56:34.016747] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:16.668 [2024-12-05 18:56:34.016942] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:16.668 [2024-12-05 18:56:34.017031] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:16.668 [2024-12-05 18:56:34.017156] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:06:17.238 18:56:34 event.event_scheduler -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:17.238 18:56:34 event.event_scheduler -- common/autotest_common.sh@868 -- # return 0 00:06:17.238 18:56:34 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:06:17.238 18:56:34 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.238 18:56:34 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:17.238 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:06:17.238 POWER: Cannot set governor of lcore 0 to userspace 00:06:17.238 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:06:17.238 POWER: Cannot set governor of lcore 0 to performance 00:06:17.238 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:06:17.238 POWER: Cannot set governor of lcore 0 to userspace 00:06:17.238 GUEST_CHANNEL: Unable to to connect to '/dev/virtio-ports/virtio.serial.port.poweragent.0' with error No such file or directory 00:06:17.238 POWER: Unable to set Power Management Environment for lcore 0 00:06:17.238 [2024-12-05 18:56:34.661580] dpdk_governor.c: 135:_init_core: *ERROR*: Failed to initialize on core0 00:06:17.238 [2024-12-05 18:56:34.661616] dpdk_governor.c: 196:_init: *ERROR*: Failed to initialize on core0 00:06:17.238 [2024-12-05 18:56:34.661644] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:06:17.238 [2024-12-05 18:56:34.661680] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:06:17.238 [2024-12-05 18:56:34.661691] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:06:17.238 [2024-12-05 18:56:34.661703] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:06:17.238 18:56:34 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.238 18:56:34 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:06:17.238 18:56:34 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.238 18:56:34 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:17.238 [2024-12-05 18:56:34.739472] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:06:17.238 18:56:34 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.238 18:56:34 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:06:17.238 18:56:34 event.event_scheduler -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:17.238 18:56:34 event.event_scheduler -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:17.238 18:56:34 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:17.238 ************************************ 00:06:17.238 START TEST scheduler_create_thread 00:06:17.238 ************************************ 00:06:17.238 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1129 -- # scheduler_create_thread 00:06:17.238 18:56:34 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:06:17.238 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.238 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:17.238 2 00:06:17.238 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.238 18:56:34 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:06:17.238 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.238 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:17.238 3 00:06:17.238 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.238 18:56:34 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:06:17.238 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.238 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:17.498 4 00:06:17.498 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.498 18:56:34 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:06:17.498 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.498 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:17.498 5 00:06:17.498 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.498 18:56:34 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:06:17.498 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.498 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:17.498 6 00:06:17.498 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.498 18:56:34 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:06:17.499 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.499 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:17.499 7 00:06:17.499 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.499 18:56:34 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:06:17.499 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.499 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:17.499 8 00:06:17.499 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.499 18:56:34 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:06:17.499 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.499 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:17.499 9 00:06:17.499 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.499 18:56:34 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:06:17.499 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.499 18:56:34 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:17.758 10 00:06:17.758 18:56:35 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:17.758 18:56:35 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:06:17.758 18:56:35 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:17.758 18:56:35 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:19.139 18:56:36 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:19.139 18:56:36 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:06:19.139 18:56:36 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:06:19.139 18:56:36 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:19.139 18:56:36 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:20.079 18:56:37 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:20.079 18:56:37 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:06:20.079 18:56:37 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:20.079 18:56:37 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:20.649 18:56:38 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:20.909 18:56:38 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:06:20.909 18:56:38 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:06:20.909 18:56:38 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:20.909 18:56:38 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:21.478 18:56:38 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:21.478 00:06:21.478 real 0m4.211s 00:06:21.478 user 0m0.023s 00:06:21.478 ************************************ 00:06:21.478 END TEST scheduler_create_thread 00:06:21.478 ************************************ 00:06:21.478 sys 0m0.013s 00:06:21.478 18:56:38 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:21.478 18:56:38 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:21.478 18:56:39 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:06:21.478 18:56:39 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 69946 00:06:21.478 18:56:39 event.event_scheduler -- common/autotest_common.sh@954 -- # '[' -z 69946 ']' 00:06:21.478 18:56:39 event.event_scheduler -- common/autotest_common.sh@958 -- # kill -0 69946 00:06:21.478 18:56:39 event.event_scheduler -- common/autotest_common.sh@959 -- # uname 00:06:21.478 18:56:39 event.event_scheduler -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:21.478 18:56:39 event.event_scheduler -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 69946 00:06:21.738 killing process with pid 69946 00:06:21.738 18:56:39 event.event_scheduler -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:06:21.738 18:56:39 event.event_scheduler -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:06:21.738 18:56:39 event.event_scheduler -- common/autotest_common.sh@972 -- # echo 'killing process with pid 69946' 00:06:21.738 18:56:39 event.event_scheduler -- common/autotest_common.sh@973 -- # kill 69946 00:06:21.738 18:56:39 event.event_scheduler -- common/autotest_common.sh@978 -- # wait 69946 00:06:21.738 [2024-12-05 18:56:39.243304] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:06:21.998 ************************************ 00:06:21.998 END TEST event_scheduler 00:06:21.998 ************************************ 00:06:21.998 00:06:21.998 real 0m5.982s 00:06:21.998 user 0m12.919s 00:06:21.998 sys 0m0.489s 00:06:21.998 18:56:39 event.event_scheduler -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:21.998 18:56:39 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:22.258 18:56:39 event -- event/event.sh@51 -- # modprobe -n nbd 00:06:22.258 18:56:39 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:06:22.258 18:56:39 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:22.258 18:56:39 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:22.258 18:56:39 event -- common/autotest_common.sh@10 -- # set +x 00:06:22.258 ************************************ 00:06:22.258 START TEST app_repeat 00:06:22.258 ************************************ 00:06:22.258 18:56:39 event.app_repeat -- common/autotest_common.sh@1129 -- # app_repeat_test 00:06:22.258 18:56:39 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:22.258 18:56:39 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:22.258 18:56:39 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:06:22.258 18:56:39 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:22.258 18:56:39 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:06:22.258 18:56:39 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:06:22.258 18:56:39 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:06:22.258 18:56:39 event.app_repeat -- event/event.sh@19 -- # repeat_pid=70058 00:06:22.258 18:56:39 event.app_repeat -- event/event.sh@18 -- # /home/vagrant/spdk_repo/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:06:22.258 18:56:39 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:06:22.258 18:56:39 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 70058' 00:06:22.258 Process app_repeat pid: 70058 00:06:22.258 spdk_app_start Round 0 00:06:22.258 18:56:39 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:22.258 18:56:39 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:06:22.259 18:56:39 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70058 /var/tmp/spdk-nbd.sock 00:06:22.259 18:56:39 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70058 ']' 00:06:22.259 18:56:39 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:22.259 18:56:39 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:22.259 18:56:39 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:22.259 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:22.259 18:56:39 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:22.259 18:56:39 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:22.259 [2024-12-05 18:56:39.645951] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:22.259 [2024-12-05 18:56:39.646136] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70058 ] 00:06:22.259 [2024-12-05 18:56:39.799373] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:22.518 [2024-12-05 18:56:39.824687] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:22.518 [2024-12-05 18:56:39.824825] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:23.122 18:56:40 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:23.122 18:56:40 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:06:23.122 18:56:40 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:23.122 Malloc0 00:06:23.400 18:56:40 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:23.400 Malloc1 00:06:23.400 18:56:40 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:23.400 18:56:40 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:23.400 18:56:40 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:23.400 18:56:40 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:23.400 18:56:40 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:23.400 18:56:40 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:23.400 18:56:40 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:23.400 18:56:40 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:23.400 18:56:40 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:23.400 18:56:40 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:23.400 18:56:40 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:23.400 18:56:40 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:23.400 18:56:40 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:23.400 18:56:40 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:23.401 18:56:40 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:23.401 18:56:40 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:23.660 /dev/nbd0 00:06:23.660 18:56:41 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:23.660 18:56:41 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:23.660 18:56:41 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:06:23.660 18:56:41 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:23.660 18:56:41 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:23.660 18:56:41 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:23.660 18:56:41 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:06:23.660 18:56:41 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:23.660 18:56:41 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:23.660 18:56:41 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:23.660 18:56:41 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:23.660 1+0 records in 00:06:23.660 1+0 records out 00:06:23.660 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000340232 s, 12.0 MB/s 00:06:23.661 18:56:41 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:23.661 18:56:41 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:23.661 18:56:41 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:23.661 18:56:41 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:23.661 18:56:41 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:23.661 18:56:41 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:23.661 18:56:41 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:23.661 18:56:41 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:23.921 /dev/nbd1 00:06:23.921 18:56:41 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:23.921 18:56:41 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:23.921 18:56:41 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:06:23.921 18:56:41 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:23.921 18:56:41 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:23.921 18:56:41 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:23.921 18:56:41 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:06:23.921 18:56:41 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:23.921 18:56:41 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:23.921 18:56:41 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:23.921 18:56:41 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:23.921 1+0 records in 00:06:23.921 1+0 records out 00:06:23.921 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000335054 s, 12.2 MB/s 00:06:23.921 18:56:41 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:23.921 18:56:41 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:23.921 18:56:41 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:23.921 18:56:41 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:23.921 18:56:41 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:23.921 18:56:41 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:23.921 18:56:41 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:23.921 18:56:41 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:23.921 18:56:41 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:23.921 18:56:41 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:24.181 { 00:06:24.181 "nbd_device": "/dev/nbd0", 00:06:24.181 "bdev_name": "Malloc0" 00:06:24.181 }, 00:06:24.181 { 00:06:24.181 "nbd_device": "/dev/nbd1", 00:06:24.181 "bdev_name": "Malloc1" 00:06:24.181 } 00:06:24.181 ]' 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:24.181 { 00:06:24.181 "nbd_device": "/dev/nbd0", 00:06:24.181 "bdev_name": "Malloc0" 00:06:24.181 }, 00:06:24.181 { 00:06:24.181 "nbd_device": "/dev/nbd1", 00:06:24.181 "bdev_name": "Malloc1" 00:06:24.181 } 00:06:24.181 ]' 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:24.181 /dev/nbd1' 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:24.181 /dev/nbd1' 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:24.181 256+0 records in 00:06:24.181 256+0 records out 00:06:24.181 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00468061 s, 224 MB/s 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:24.181 256+0 records in 00:06:24.181 256+0 records out 00:06:24.181 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0234224 s, 44.8 MB/s 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:24.181 256+0 records in 00:06:24.181 256+0 records out 00:06:24.181 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0225791 s, 46.4 MB/s 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:24.181 18:56:41 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:24.441 18:56:41 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:24.441 18:56:41 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:24.441 18:56:41 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:24.441 18:56:41 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:24.441 18:56:41 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:24.441 18:56:41 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:24.441 18:56:41 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:24.441 18:56:41 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:24.441 18:56:41 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:24.441 18:56:41 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:24.701 18:56:42 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:24.701 18:56:42 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:24.701 18:56:42 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:24.701 18:56:42 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:24.701 18:56:42 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:24.701 18:56:42 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:24.701 18:56:42 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:24.701 18:56:42 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:24.701 18:56:42 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:24.701 18:56:42 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:24.701 18:56:42 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:24.960 18:56:42 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:24.960 18:56:42 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:24.960 18:56:42 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:24.960 18:56:42 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:24.960 18:56:42 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:24.960 18:56:42 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:24.960 18:56:42 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:24.960 18:56:42 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:24.960 18:56:42 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:24.960 18:56:42 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:24.960 18:56:42 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:24.960 18:56:42 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:24.960 18:56:42 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:25.219 18:56:42 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:25.219 [2024-12-05 18:56:42.726183] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:25.219 [2024-12-05 18:56:42.748936] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:25.219 [2024-12-05 18:56:42.748940] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:25.478 [2024-12-05 18:56:42.790551] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:25.478 [2024-12-05 18:56:42.790649] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:28.770 spdk_app_start Round 1 00:06:28.770 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:28.770 18:56:45 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:28.770 18:56:45 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:06:28.770 18:56:45 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70058 /var/tmp/spdk-nbd.sock 00:06:28.770 18:56:45 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70058 ']' 00:06:28.770 18:56:45 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:28.770 18:56:45 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:28.770 18:56:45 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:28.770 18:56:45 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:28.770 18:56:45 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:28.770 18:56:45 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:28.770 18:56:45 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:06:28.770 18:56:45 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:28.770 Malloc0 00:06:28.770 18:56:46 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:28.770 Malloc1 00:06:28.770 18:56:46 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:28.770 18:56:46 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:28.770 18:56:46 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:28.770 18:56:46 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:28.770 18:56:46 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:28.770 18:56:46 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:28.770 18:56:46 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:28.770 18:56:46 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:28.770 18:56:46 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:28.770 18:56:46 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:28.770 18:56:46 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:28.770 18:56:46 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:28.770 18:56:46 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:28.770 18:56:46 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:28.770 18:56:46 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:28.770 18:56:46 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:29.030 /dev/nbd0 00:06:29.030 18:56:46 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:29.030 18:56:46 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:29.030 18:56:46 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:06:29.030 18:56:46 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:29.030 18:56:46 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:29.030 18:56:46 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:29.030 18:56:46 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:06:29.030 18:56:46 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:29.030 18:56:46 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:29.030 18:56:46 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:29.030 18:56:46 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:29.030 1+0 records in 00:06:29.030 1+0 records out 00:06:29.030 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000445642 s, 9.2 MB/s 00:06:29.030 18:56:46 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:29.030 18:56:46 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:29.030 18:56:46 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:29.030 18:56:46 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:29.030 18:56:46 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:29.030 18:56:46 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:29.030 18:56:46 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:29.030 18:56:46 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:29.289 /dev/nbd1 00:06:29.289 18:56:46 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:29.289 18:56:46 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:29.289 18:56:46 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:06:29.289 18:56:46 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:29.289 18:56:46 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:29.289 18:56:46 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:29.289 18:56:46 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:06:29.289 18:56:46 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:29.289 18:56:46 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:29.289 18:56:46 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:29.290 18:56:46 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:29.290 1+0 records in 00:06:29.290 1+0 records out 00:06:29.290 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000379151 s, 10.8 MB/s 00:06:29.290 18:56:46 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:29.290 18:56:46 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:29.290 18:56:46 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:29.290 18:56:46 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:29.290 18:56:46 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:29.290 18:56:46 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:29.290 18:56:46 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:29.290 18:56:46 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:29.290 18:56:46 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:29.290 18:56:46 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:29.549 18:56:46 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:29.549 { 00:06:29.549 "nbd_device": "/dev/nbd0", 00:06:29.549 "bdev_name": "Malloc0" 00:06:29.549 }, 00:06:29.549 { 00:06:29.549 "nbd_device": "/dev/nbd1", 00:06:29.549 "bdev_name": "Malloc1" 00:06:29.549 } 00:06:29.549 ]' 00:06:29.549 18:56:46 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:29.549 { 00:06:29.549 "nbd_device": "/dev/nbd0", 00:06:29.549 "bdev_name": "Malloc0" 00:06:29.549 }, 00:06:29.549 { 00:06:29.549 "nbd_device": "/dev/nbd1", 00:06:29.549 "bdev_name": "Malloc1" 00:06:29.549 } 00:06:29.549 ]' 00:06:29.549 18:56:46 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:29.549 18:56:46 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:29.549 /dev/nbd1' 00:06:29.549 18:56:46 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:29.549 /dev/nbd1' 00:06:29.549 18:56:46 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:29.549 18:56:46 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:29.549 18:56:46 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:29.549 18:56:46 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:29.549 18:56:46 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:29.549 18:56:46 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:29.549 18:56:46 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:29.550 18:56:46 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:29.550 18:56:46 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:29.550 18:56:46 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:29.550 18:56:46 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:29.550 18:56:46 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:29.550 256+0 records in 00:06:29.550 256+0 records out 00:06:29.550 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00506686 s, 207 MB/s 00:06:29.550 18:56:46 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:29.550 18:56:46 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:29.550 256+0 records in 00:06:29.550 256+0 records out 00:06:29.550 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0242788 s, 43.2 MB/s 00:06:29.550 18:56:46 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:29.550 18:56:46 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:29.550 256+0 records in 00:06:29.550 256+0 records out 00:06:29.550 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0252829 s, 41.5 MB/s 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:29.550 18:56:47 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:29.810 18:56:47 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:29.810 18:56:47 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:29.810 18:56:47 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:29.810 18:56:47 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:29.810 18:56:47 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:29.810 18:56:47 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:29.810 18:56:47 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:29.810 18:56:47 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:29.810 18:56:47 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:29.810 18:56:47 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:30.069 18:56:47 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:30.069 18:56:47 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:30.069 18:56:47 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:30.069 18:56:47 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:30.069 18:56:47 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:30.069 18:56:47 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:30.069 18:56:47 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:30.069 18:56:47 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:30.069 18:56:47 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:30.069 18:56:47 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:30.069 18:56:47 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:30.329 18:56:47 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:30.329 18:56:47 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:30.329 18:56:47 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:30.329 18:56:47 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:30.329 18:56:47 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:30.329 18:56:47 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:30.329 18:56:47 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:30.329 18:56:47 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:30.329 18:56:47 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:30.329 18:56:47 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:30.329 18:56:47 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:30.329 18:56:47 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:30.329 18:56:47 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:30.588 18:56:47 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:30.588 [2024-12-05 18:56:48.072127] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:30.588 [2024-12-05 18:56:48.096521] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:30.588 [2024-12-05 18:56:48.096538] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:30.588 [2024-12-05 18:56:48.138869] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:30.588 [2024-12-05 18:56:48.138933] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:33.881 spdk_app_start Round 2 00:06:33.881 18:56:50 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:33.881 18:56:50 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:06:33.881 18:56:50 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70058 /var/tmp/spdk-nbd.sock 00:06:33.881 18:56:50 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70058 ']' 00:06:33.881 18:56:50 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:33.881 18:56:50 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:33.881 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:33.881 18:56:50 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:33.881 18:56:50 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:33.881 18:56:50 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:33.881 18:56:51 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:33.881 18:56:51 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:06:33.881 18:56:51 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:33.881 Malloc0 00:06:33.881 18:56:51 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:34.141 Malloc1 00:06:34.141 18:56:51 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:34.141 18:56:51 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:34.141 18:56:51 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:34.141 18:56:51 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:34.141 18:56:51 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:34.141 18:56:51 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:34.141 18:56:51 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:34.141 18:56:51 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:34.141 18:56:51 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:34.141 18:56:51 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:34.141 18:56:51 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:34.141 18:56:51 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:34.141 18:56:51 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:34.141 18:56:51 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:34.141 18:56:51 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:34.141 18:56:51 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:34.401 /dev/nbd0 00:06:34.401 18:56:51 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:34.401 18:56:51 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:34.401 18:56:51 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:06:34.401 18:56:51 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:34.401 18:56:51 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:34.401 18:56:51 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:34.401 18:56:51 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:06:34.401 18:56:51 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:34.401 18:56:51 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:34.401 18:56:51 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:34.401 18:56:51 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:34.401 1+0 records in 00:06:34.401 1+0 records out 00:06:34.401 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000401385 s, 10.2 MB/s 00:06:34.401 18:56:51 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:34.401 18:56:51 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:34.401 18:56:51 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:34.402 18:56:51 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:34.402 18:56:51 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:34.402 18:56:51 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:34.402 18:56:51 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:34.402 18:56:51 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:34.662 /dev/nbd1 00:06:34.662 18:56:52 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:34.662 18:56:52 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:34.662 18:56:52 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:06:34.662 18:56:52 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:06:34.662 18:56:52 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:06:34.662 18:56:52 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:06:34.662 18:56:52 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:06:34.662 18:56:52 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:06:34.662 18:56:52 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:06:34.662 18:56:52 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:06:34.662 18:56:52 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:34.662 1+0 records in 00:06:34.662 1+0 records out 00:06:34.662 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000335517 s, 12.2 MB/s 00:06:34.662 18:56:52 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:34.662 18:56:52 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:06:34.662 18:56:52 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:34.662 18:56:52 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:06:34.662 18:56:52 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:06:34.662 18:56:52 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:34.662 18:56:52 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:34.662 18:56:52 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:34.662 18:56:52 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:34.662 18:56:52 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:34.923 { 00:06:34.923 "nbd_device": "/dev/nbd0", 00:06:34.923 "bdev_name": "Malloc0" 00:06:34.923 }, 00:06:34.923 { 00:06:34.923 "nbd_device": "/dev/nbd1", 00:06:34.923 "bdev_name": "Malloc1" 00:06:34.923 } 00:06:34.923 ]' 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:34.923 { 00:06:34.923 "nbd_device": "/dev/nbd0", 00:06:34.923 "bdev_name": "Malloc0" 00:06:34.923 }, 00:06:34.923 { 00:06:34.923 "nbd_device": "/dev/nbd1", 00:06:34.923 "bdev_name": "Malloc1" 00:06:34.923 } 00:06:34.923 ]' 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:34.923 /dev/nbd1' 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:34.923 /dev/nbd1' 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:34.923 18:56:52 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:34.923 256+0 records in 00:06:34.923 256+0 records out 00:06:34.923 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0132014 s, 79.4 MB/s 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:34.924 256+0 records in 00:06:34.924 256+0 records out 00:06:34.924 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0201738 s, 52.0 MB/s 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:34.924 256+0 records in 00:06:34.924 256+0 records out 00:06:34.924 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0202005 s, 51.9 MB/s 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:34.924 18:56:52 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:35.183 18:56:52 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:35.183 18:56:52 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:35.183 18:56:52 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:35.183 18:56:52 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:35.183 18:56:52 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:35.183 18:56:52 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:35.183 18:56:52 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:35.183 18:56:52 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:35.183 18:56:52 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:35.183 18:56:52 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:35.443 18:56:52 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:35.443 18:56:52 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:35.443 18:56:52 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:35.443 18:56:52 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:35.443 18:56:52 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:35.443 18:56:52 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:35.443 18:56:52 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:35.443 18:56:52 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:35.443 18:56:52 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:35.443 18:56:52 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:35.443 18:56:52 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:35.702 18:56:53 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:35.702 18:56:53 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:35.702 18:56:53 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:35.702 18:56:53 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:35.702 18:56:53 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:35.702 18:56:53 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:35.702 18:56:53 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:35.702 18:56:53 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:35.702 18:56:53 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:35.702 18:56:53 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:35.702 18:56:53 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:35.702 18:56:53 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:35.702 18:56:53 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:35.961 18:56:53 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:35.961 [2024-12-05 18:56:53.437569] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:35.961 [2024-12-05 18:56:53.460239] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:35.961 [2024-12-05 18:56:53.460240] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:35.961 [2024-12-05 18:56:53.501848] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:35.961 [2024-12-05 18:56:53.501930] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:39.251 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:39.251 18:56:56 event.app_repeat -- event/event.sh@38 -- # waitforlisten 70058 /var/tmp/spdk-nbd.sock 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 70058 ']' 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:06:39.251 18:56:56 event.app_repeat -- event/event.sh@39 -- # killprocess 70058 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@954 -- # '[' -z 70058 ']' 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@958 -- # kill -0 70058 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@959 -- # uname 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70058 00:06:39.251 killing process with pid 70058 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70058' 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@973 -- # kill 70058 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@978 -- # wait 70058 00:06:39.251 spdk_app_start is called in Round 0. 00:06:39.251 Shutdown signal received, stop current app iteration 00:06:39.251 Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 reinitialization... 00:06:39.251 spdk_app_start is called in Round 1. 00:06:39.251 Shutdown signal received, stop current app iteration 00:06:39.251 Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 reinitialization... 00:06:39.251 spdk_app_start is called in Round 2. 00:06:39.251 Shutdown signal received, stop current app iteration 00:06:39.251 Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 reinitialization... 00:06:39.251 spdk_app_start is called in Round 3. 00:06:39.251 Shutdown signal received, stop current app iteration 00:06:39.251 18:56:56 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:06:39.251 18:56:56 event.app_repeat -- event/event.sh@42 -- # return 0 00:06:39.251 00:06:39.251 real 0m17.160s 00:06:39.251 user 0m37.888s 00:06:39.251 sys 0m2.590s 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:39.251 18:56:56 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:39.251 ************************************ 00:06:39.251 END TEST app_repeat 00:06:39.251 ************************************ 00:06:39.251 18:56:56 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:06:39.251 18:56:56 event -- event/event.sh@55 -- # run_test cpu_locks /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:06:39.251 18:56:56 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:39.251 18:56:56 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:39.251 18:56:56 event -- common/autotest_common.sh@10 -- # set +x 00:06:39.520 ************************************ 00:06:39.520 START TEST cpu_locks 00:06:39.520 ************************************ 00:06:39.520 18:56:56 event.cpu_locks -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:06:39.520 * Looking for test storage... 00:06:39.520 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:06:39.520 18:56:56 event.cpu_locks -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:06:39.520 18:56:56 event.cpu_locks -- common/autotest_common.sh@1711 -- # lcov --version 00:06:39.520 18:56:56 event.cpu_locks -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:06:39.520 18:56:57 event.cpu_locks -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:39.520 18:56:57 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:06:39.520 18:56:57 event.cpu_locks -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:39.520 18:56:57 event.cpu_locks -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:06:39.520 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:39.520 --rc genhtml_branch_coverage=1 00:06:39.520 --rc genhtml_function_coverage=1 00:06:39.520 --rc genhtml_legend=1 00:06:39.520 --rc geninfo_all_blocks=1 00:06:39.520 --rc geninfo_unexecuted_blocks=1 00:06:39.520 00:06:39.520 ' 00:06:39.520 18:56:57 event.cpu_locks -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:06:39.520 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:39.520 --rc genhtml_branch_coverage=1 00:06:39.520 --rc genhtml_function_coverage=1 00:06:39.520 --rc genhtml_legend=1 00:06:39.520 --rc geninfo_all_blocks=1 00:06:39.520 --rc geninfo_unexecuted_blocks=1 00:06:39.520 00:06:39.520 ' 00:06:39.520 18:56:57 event.cpu_locks -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:06:39.520 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:39.520 --rc genhtml_branch_coverage=1 00:06:39.520 --rc genhtml_function_coverage=1 00:06:39.520 --rc genhtml_legend=1 00:06:39.520 --rc geninfo_all_blocks=1 00:06:39.520 --rc geninfo_unexecuted_blocks=1 00:06:39.520 00:06:39.520 ' 00:06:39.520 18:56:57 event.cpu_locks -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:06:39.520 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:39.520 --rc genhtml_branch_coverage=1 00:06:39.520 --rc genhtml_function_coverage=1 00:06:39.520 --rc genhtml_legend=1 00:06:39.520 --rc geninfo_all_blocks=1 00:06:39.520 --rc geninfo_unexecuted_blocks=1 00:06:39.520 00:06:39.520 ' 00:06:39.520 18:56:57 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:06:39.520 18:56:57 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:06:39.520 18:56:57 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:06:39.520 18:56:57 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:06:39.520 18:56:57 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:39.520 18:56:57 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:39.520 18:56:57 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:39.520 ************************************ 00:06:39.520 START TEST default_locks 00:06:39.520 ************************************ 00:06:39.520 18:56:57 event.cpu_locks.default_locks -- common/autotest_common.sh@1129 -- # default_locks 00:06:39.520 18:56:57 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=70483 00:06:39.520 18:56:57 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:39.520 18:56:57 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 70483 00:06:39.520 18:56:57 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 70483 ']' 00:06:39.520 18:56:57 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:39.520 18:56:57 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:39.520 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:39.520 18:56:57 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:39.520 18:56:57 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:39.520 18:56:57 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:39.792 [2024-12-05 18:56:57.172044] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:39.792 [2024-12-05 18:56:57.172244] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70483 ] 00:06:39.792 [2024-12-05 18:56:57.348399] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:40.052 [2024-12-05 18:56:57.372555] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:40.622 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:40.622 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 0 00:06:40.622 18:56:58 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 70483 00:06:40.622 18:56:58 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 70483 00:06:40.622 18:56:58 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:40.882 18:56:58 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 70483 00:06:40.882 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # '[' -z 70483 ']' 00:06:40.882 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@958 -- # kill -0 70483 00:06:40.882 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # uname 00:06:40.882 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:40.882 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70483 00:06:40.882 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:40.882 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:40.882 killing process with pid 70483 00:06:40.882 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70483' 00:06:40.882 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@973 -- # kill 70483 00:06:40.882 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@978 -- # wait 70483 00:06:41.142 18:56:58 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 70483 00:06:41.142 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # local es=0 00:06:41.142 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 70483 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # waitforlisten 70483 00:06:41.143 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:41.143 ERROR: process (pid: 70483) is no longer running 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 70483 ']' 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:41.143 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (70483) - No such process 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 1 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # es=1 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:06:41.143 00:06:41.143 real 0m1.534s 00:06:41.143 user 0m1.511s 00:06:41.143 sys 0m0.527s 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:41.143 18:56:58 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:41.143 ************************************ 00:06:41.143 END TEST default_locks 00:06:41.143 ************************************ 00:06:41.143 18:56:58 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:06:41.143 18:56:58 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:41.143 18:56:58 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:41.143 18:56:58 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:41.143 ************************************ 00:06:41.143 START TEST default_locks_via_rpc 00:06:41.143 ************************************ 00:06:41.143 18:56:58 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1129 -- # default_locks_via_rpc 00:06:41.143 18:56:58 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=70530 00:06:41.143 18:56:58 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:41.143 18:56:58 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 70530 00:06:41.143 18:56:58 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70530 ']' 00:06:41.143 18:56:58 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:41.143 18:56:58 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:41.143 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:41.143 18:56:58 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:41.143 18:56:58 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:41.143 18:56:58 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:41.403 [2024-12-05 18:56:58.754919] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:41.403 [2024-12-05 18:56:58.755062] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70530 ] 00:06:41.403 [2024-12-05 18:56:58.911776] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:41.403 [2024-12-05 18:56:58.936182] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 70530 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:42.343 18:56:59 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 70530 00:06:42.603 18:56:59 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 70530 00:06:42.603 18:56:59 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # '[' -z 70530 ']' 00:06:42.603 18:56:59 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@958 -- # kill -0 70530 00:06:42.603 18:56:59 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # uname 00:06:42.603 18:56:59 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:42.603 18:56:59 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70530 00:06:42.603 18:57:00 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:42.603 18:57:00 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:42.603 killing process with pid 70530 00:06:42.603 18:57:00 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70530' 00:06:42.603 18:57:00 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@973 -- # kill 70530 00:06:42.603 18:57:00 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@978 -- # wait 70530 00:06:42.862 00:06:42.862 real 0m1.727s 00:06:42.862 user 0m1.720s 00:06:42.862 sys 0m0.573s 00:06:42.862 18:57:00 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:42.862 18:57:00 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:42.862 ************************************ 00:06:42.862 END TEST default_locks_via_rpc 00:06:42.862 ************************************ 00:06:43.122 18:57:00 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:06:43.122 18:57:00 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:43.122 18:57:00 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:43.122 18:57:00 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:43.122 ************************************ 00:06:43.122 START TEST non_locking_app_on_locked_coremask 00:06:43.122 ************************************ 00:06:43.122 18:57:00 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # non_locking_app_on_locked_coremask 00:06:43.122 18:57:00 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=70577 00:06:43.122 18:57:00 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:43.122 18:57:00 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 70577 /var/tmp/spdk.sock 00:06:43.122 18:57:00 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70577 ']' 00:06:43.122 18:57:00 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:43.122 18:57:00 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:43.122 18:57:00 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:43.122 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:43.122 18:57:00 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:43.122 18:57:00 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:43.122 [2024-12-05 18:57:00.554027] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:43.122 [2024-12-05 18:57:00.554203] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70577 ] 00:06:43.382 [2024-12-05 18:57:00.710761] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:43.382 [2024-12-05 18:57:00.736555] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:43.952 18:57:01 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:43.952 18:57:01 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:43.952 18:57:01 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:06:43.952 18:57:01 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=70593 00:06:43.952 18:57:01 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 70593 /var/tmp/spdk2.sock 00:06:43.952 18:57:01 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70593 ']' 00:06:43.952 18:57:01 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:43.952 18:57:01 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:43.952 18:57:01 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:43.952 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:43.952 18:57:01 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:43.952 18:57:01 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:43.952 [2024-12-05 18:57:01.415089] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:43.952 [2024-12-05 18:57:01.415249] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70593 ] 00:06:44.212 [2024-12-05 18:57:01.563358] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:44.212 [2024-12-05 18:57:01.563412] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:44.212 [2024-12-05 18:57:01.611843] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:44.782 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:44.782 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:44.782 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 70577 00:06:44.782 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70577 00:06:44.782 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:45.351 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 70577 00:06:45.351 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70577 ']' 00:06:45.351 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 70577 00:06:45.351 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:06:45.351 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:45.351 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70577 00:06:45.351 killing process with pid 70577 00:06:45.351 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:45.351 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:45.351 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70577' 00:06:45.351 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 70577 00:06:45.351 18:57:02 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 70577 00:06:46.289 18:57:03 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 70593 00:06:46.289 18:57:03 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70593 ']' 00:06:46.289 18:57:03 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 70593 00:06:46.289 18:57:03 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:06:46.289 18:57:03 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:46.289 18:57:03 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70593 00:06:46.289 killing process with pid 70593 00:06:46.289 18:57:03 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:46.289 18:57:03 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:46.289 18:57:03 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70593' 00:06:46.289 18:57:03 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 70593 00:06:46.289 18:57:03 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 70593 00:06:46.549 00:06:46.549 real 0m3.466s 00:06:46.549 user 0m3.618s 00:06:46.549 sys 0m1.051s 00:06:46.549 18:57:03 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:46.549 18:57:03 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:46.549 ************************************ 00:06:46.549 END TEST non_locking_app_on_locked_coremask 00:06:46.549 ************************************ 00:06:46.549 18:57:03 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:06:46.549 18:57:03 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:46.549 18:57:03 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:46.549 18:57:03 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:46.549 ************************************ 00:06:46.549 START TEST locking_app_on_unlocked_coremask 00:06:46.549 ************************************ 00:06:46.549 18:57:03 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_unlocked_coremask 00:06:46.549 18:57:03 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=70658 00:06:46.549 18:57:03 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:06:46.549 18:57:03 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 70658 /var/tmp/spdk.sock 00:06:46.549 18:57:03 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70658 ']' 00:06:46.549 18:57:03 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:46.549 18:57:03 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:46.549 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:46.549 18:57:03 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:46.549 18:57:03 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:46.549 18:57:03 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:46.549 [2024-12-05 18:57:04.081536] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:46.549 [2024-12-05 18:57:04.081684] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70658 ] 00:06:46.810 [2024-12-05 18:57:04.235193] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:46.810 [2024-12-05 18:57:04.235244] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:46.810 [2024-12-05 18:57:04.259644] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:47.380 18:57:04 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:47.380 18:57:04 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:47.380 18:57:04 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:06:47.380 18:57:04 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=70669 00:06:47.380 18:57:04 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 70669 /var/tmp/spdk2.sock 00:06:47.380 18:57:04 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70669 ']' 00:06:47.380 18:57:04 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:47.380 18:57:04 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:47.380 18:57:04 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:47.380 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:47.380 18:57:04 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:47.380 18:57:04 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:47.639 [2024-12-05 18:57:04.940722] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:47.639 [2024-12-05 18:57:04.940852] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70669 ] 00:06:47.639 [2024-12-05 18:57:05.092952] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:47.639 [2024-12-05 18:57:05.141498] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:48.575 18:57:05 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:48.575 18:57:05 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:48.575 18:57:05 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 70669 00:06:48.575 18:57:05 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70669 00:06:48.575 18:57:05 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:49.158 18:57:06 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 70658 00:06:49.159 18:57:06 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70658 ']' 00:06:49.159 18:57:06 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 70658 00:06:49.159 18:57:06 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:06:49.159 18:57:06 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:49.159 18:57:06 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70658 00:06:49.159 18:57:06 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:49.159 18:57:06 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:49.159 killing process with pid 70658 00:06:49.159 18:57:06 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70658' 00:06:49.159 18:57:06 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 70658 00:06:49.159 18:57:06 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 70658 00:06:50.096 18:57:07 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 70669 00:06:50.096 18:57:07 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70669 ']' 00:06:50.096 18:57:07 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 70669 00:06:50.097 18:57:07 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:06:50.097 18:57:07 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:50.097 18:57:07 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70669 00:06:50.097 18:57:07 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:50.097 18:57:07 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:50.097 killing process with pid 70669 00:06:50.097 18:57:07 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70669' 00:06:50.097 18:57:07 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 70669 00:06:50.097 18:57:07 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 70669 00:06:50.356 00:06:50.356 real 0m3.704s 00:06:50.356 user 0m3.879s 00:06:50.356 sys 0m1.143s 00:06:50.356 18:57:07 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:50.356 18:57:07 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:50.356 ************************************ 00:06:50.356 END TEST locking_app_on_unlocked_coremask 00:06:50.356 ************************************ 00:06:50.356 18:57:07 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:06:50.356 18:57:07 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:50.356 18:57:07 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:50.356 18:57:07 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:50.356 ************************************ 00:06:50.356 START TEST locking_app_on_locked_coremask 00:06:50.356 ************************************ 00:06:50.356 18:57:07 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_locked_coremask 00:06:50.356 18:57:07 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=70738 00:06:50.356 18:57:07 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:50.356 18:57:07 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 70738 /var/tmp/spdk.sock 00:06:50.356 18:57:07 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70738 ']' 00:06:50.356 18:57:07 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:50.356 18:57:07 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:50.356 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:50.356 18:57:07 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:50.356 18:57:07 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:50.356 18:57:07 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:50.356 [2024-12-05 18:57:07.859628] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:50.357 [2024-12-05 18:57:07.859779] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70738 ] 00:06:50.615 [2024-12-05 18:57:08.013494] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:50.615 [2024-12-05 18:57:08.037692] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=70754 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 70754 /var/tmp/spdk2.sock 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # local es=0 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 70754 /var/tmp/spdk2.sock 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # waitforlisten 70754 /var/tmp/spdk2.sock 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 70754 ']' 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:51.183 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:51.183 18:57:08 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:51.443 [2024-12-05 18:57:08.742511] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:51.443 [2024-12-05 18:57:08.742637] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70754 ] 00:06:51.443 [2024-12-05 18:57:08.892042] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 70738 has claimed it. 00:06:51.443 [2024-12-05 18:57:08.892101] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:06:52.013 ERROR: process (pid: 70754) is no longer running 00:06:52.013 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (70754) - No such process 00:06:52.013 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:52.013 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 1 00:06:52.013 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # es=1 00:06:52.013 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:52.013 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:52.013 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:52.013 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 70738 00:06:52.013 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70738 00:06:52.013 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:52.274 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 70738 00:06:52.274 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 70738 ']' 00:06:52.274 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 70738 00:06:52.274 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:06:52.274 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:52.274 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70738 00:06:52.274 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:52.274 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:52.274 killing process with pid 70738 00:06:52.274 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70738' 00:06:52.274 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 70738 00:06:52.274 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 70738 00:06:52.534 00:06:52.534 real 0m2.214s 00:06:52.534 user 0m2.367s 00:06:52.534 sys 0m0.630s 00:06:52.534 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:52.534 18:57:09 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:52.534 ************************************ 00:06:52.534 END TEST locking_app_on_locked_coremask 00:06:52.534 ************************************ 00:06:52.534 18:57:10 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:06:52.534 18:57:10 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:52.534 18:57:10 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:52.534 18:57:10 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:52.534 ************************************ 00:06:52.534 START TEST locking_overlapped_coremask 00:06:52.534 ************************************ 00:06:52.534 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask 00:06:52.534 18:57:10 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=70796 00:06:52.534 18:57:10 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 00:06:52.534 18:57:10 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 70796 /var/tmp/spdk.sock 00:06:52.534 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 70796 ']' 00:06:52.534 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:52.534 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:52.534 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:52.534 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:52.534 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:52.534 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:52.794 [2024-12-05 18:57:10.144645] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:52.794 [2024-12-05 18:57:10.144773] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70796 ] 00:06:52.794 [2024-12-05 18:57:10.303009] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:52.794 [2024-12-05 18:57:10.330182] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:52.794 [2024-12-05 18:57:10.330272] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:52.794 [2024-12-05 18:57:10.330425] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 0 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=70814 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 70814 /var/tmp/spdk2.sock 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # local es=0 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 70814 /var/tmp/spdk2.sock 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # waitforlisten 70814 /var/tmp/spdk2.sock 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 70814 ']' 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:53.767 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:53.767 18:57:10 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:53.767 [2024-12-05 18:57:11.053245] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:53.767 [2024-12-05 18:57:11.053376] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70814 ] 00:06:53.767 [2024-12-05 18:57:11.204233] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 70796 has claimed it. 00:06:53.767 [2024-12-05 18:57:11.204296] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:06:54.337 ERROR: process (pid: 70814) is no longer running 00:06:54.337 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 850: kill: (70814) - No such process 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 1 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # es=1 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 70796 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # '[' -z 70796 ']' 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@958 -- # kill -0 70796 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # uname 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:54.337 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70796 00:06:54.338 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:54.338 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:54.338 killing process with pid 70796 00:06:54.338 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70796' 00:06:54.338 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@973 -- # kill 70796 00:06:54.338 18:57:11 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@978 -- # wait 70796 00:06:54.598 00:06:54.598 real 0m2.045s 00:06:54.598 user 0m5.503s 00:06:54.598 sys 0m0.517s 00:06:54.598 18:57:12 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:54.598 18:57:12 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:54.598 ************************************ 00:06:54.598 END TEST locking_overlapped_coremask 00:06:54.598 ************************************ 00:06:54.598 18:57:12 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:06:54.598 18:57:12 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:54.598 18:57:12 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:54.598 18:57:12 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:54.858 ************************************ 00:06:54.858 START TEST locking_overlapped_coremask_via_rpc 00:06:54.858 ************************************ 00:06:54.858 18:57:12 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask_via_rpc 00:06:54.858 18:57:12 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=70856 00:06:54.858 18:57:12 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:06:54.858 18:57:12 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 70856 /var/tmp/spdk.sock 00:06:54.858 18:57:12 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70856 ']' 00:06:54.858 18:57:12 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:54.858 18:57:12 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:54.858 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:54.858 18:57:12 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:54.858 18:57:12 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:54.858 18:57:12 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:54.858 [2024-12-05 18:57:12.251856] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:54.858 [2024-12-05 18:57:12.252063] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70856 ] 00:06:54.858 [2024-12-05 18:57:12.408628] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:54.858 [2024-12-05 18:57:12.408686] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:55.118 [2024-12-05 18:57:12.436639] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:06:55.118 [2024-12-05 18:57:12.436763] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:55.118 [2024-12-05 18:57:12.436865] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:55.688 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:55.688 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:55.688 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:06:55.688 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=70874 00:06:55.688 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 70874 /var/tmp/spdk2.sock 00:06:55.688 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70874 ']' 00:06:55.688 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:55.688 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:55.688 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:55.688 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:55.688 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:55.688 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:55.688 [2024-12-05 18:57:13.135417] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:55.689 [2024-12-05 18:57:13.135561] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70874 ] 00:06:55.948 [2024-12-05 18:57:13.286462] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:55.948 [2024-12-05 18:57:13.286552] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:55.949 [2024-12-05 18:57:13.351127] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:06:55.949 [2024-12-05 18:57:13.351203] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:06:55.949 [2024-12-05 18:57:13.351329] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # local es=0 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:06:56.519 18:57:13 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:56.519 [2024-12-05 18:57:14.002824] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 70856 has claimed it. 00:06:56.519 request: 00:06:56.519 { 00:06:56.519 "method": "framework_enable_cpumask_locks", 00:06:56.519 "req_id": 1 00:06:56.519 } 00:06:56.519 Got JSON-RPC error response 00:06:56.519 response: 00:06:56.519 { 00:06:56.519 "code": -32603, 00:06:56.519 "message": "Failed to claim CPU core: 2" 00:06:56.519 } 00:06:56.519 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:06:56.519 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # es=1 00:06:56.519 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:06:56.519 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:06:56.519 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:06:56.519 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 70856 /var/tmp/spdk.sock 00:06:56.519 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70856 ']' 00:06:56.519 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:56.520 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:56.520 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:56.520 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:56.520 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:56.520 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:56.780 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:56.780 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:56.780 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 70874 /var/tmp/spdk2.sock 00:06:56.780 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 70874 ']' 00:06:56.780 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:56.780 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:06:56.780 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:56.780 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:56.780 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:06:56.780 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:57.039 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:06:57.039 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:06:57.040 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:06:57.040 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:06:57.040 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:06:57.040 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:06:57.040 00:06:57.040 real 0m2.285s 00:06:57.040 user 0m1.032s 00:06:57.040 sys 0m0.176s 00:06:57.040 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:57.040 18:57:14 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:57.040 ************************************ 00:06:57.040 END TEST locking_overlapped_coremask_via_rpc 00:06:57.040 ************************************ 00:06:57.040 18:57:14 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:06:57.040 18:57:14 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 70856 ]] 00:06:57.040 18:57:14 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 70856 00:06:57.040 18:57:14 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 70856 ']' 00:06:57.040 18:57:14 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 70856 00:06:57.040 18:57:14 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:06:57.040 18:57:14 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:57.040 18:57:14 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70856 00:06:57.040 killing process with pid 70856 00:06:57.040 18:57:14 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:06:57.040 18:57:14 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:06:57.040 18:57:14 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70856' 00:06:57.040 18:57:14 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 70856 00:06:57.040 18:57:14 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 70856 00:06:57.610 18:57:14 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 70874 ]] 00:06:57.610 18:57:14 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 70874 00:06:57.610 18:57:14 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 70874 ']' 00:06:57.610 18:57:14 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 70874 00:06:57.610 18:57:14 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:06:57.610 18:57:14 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:06:57.610 18:57:14 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 70874 00:06:57.610 killing process with pid 70874 00:06:57.610 18:57:14 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:06:57.610 18:57:14 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:06:57.610 18:57:14 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 70874' 00:06:57.610 18:57:14 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 70874 00:06:57.610 18:57:14 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 70874 00:06:57.869 18:57:15 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:06:57.869 18:57:15 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:06:57.869 Process with pid 70856 is not found 00:06:57.869 18:57:15 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 70856 ]] 00:06:57.869 18:57:15 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 70856 00:06:57.869 18:57:15 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 70856 ']' 00:06:57.869 18:57:15 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 70856 00:06:57.869 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (70856) - No such process 00:06:57.869 18:57:15 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 70856 is not found' 00:06:57.869 18:57:15 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 70874 ]] 00:06:57.869 18:57:15 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 70874 00:06:57.869 18:57:15 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 70874 ']' 00:06:57.869 18:57:15 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 70874 00:06:57.869 Process with pid 70874 is not found 00:06:57.869 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (70874) - No such process 00:06:57.869 18:57:15 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 70874 is not found' 00:06:57.869 18:57:15 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:06:57.869 00:06:57.869 real 0m18.516s 00:06:57.869 user 0m30.990s 00:06:57.869 sys 0m5.723s 00:06:57.869 18:57:15 event.cpu_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:57.869 18:57:15 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:57.869 ************************************ 00:06:57.869 END TEST cpu_locks 00:06:57.869 ************************************ 00:06:57.869 ************************************ 00:06:57.869 END TEST event 00:06:57.869 ************************************ 00:06:57.869 00:06:57.869 real 0m46.140s 00:06:57.869 user 1m28.293s 00:06:57.869 sys 0m9.468s 00:06:57.869 18:57:15 event -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:57.869 18:57:15 event -- common/autotest_common.sh@10 -- # set +x 00:06:58.128 18:57:15 -- spdk/autotest.sh@169 -- # run_test thread /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:06:58.128 18:57:15 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:06:58.128 18:57:15 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:58.128 18:57:15 -- common/autotest_common.sh@10 -- # set +x 00:06:58.128 ************************************ 00:06:58.128 START TEST thread 00:06:58.128 ************************************ 00:06:58.128 18:57:15 thread -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:06:58.128 * Looking for test storage... 00:06:58.128 * Found test storage at /home/vagrant/spdk_repo/spdk/test/thread 00:06:58.128 18:57:15 thread -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:06:58.128 18:57:15 thread -- common/autotest_common.sh@1711 -- # lcov --version 00:06:58.128 18:57:15 thread -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:06:58.128 18:57:15 thread -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:06:58.128 18:57:15 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:58.128 18:57:15 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:58.128 18:57:15 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:58.128 18:57:15 thread -- scripts/common.sh@336 -- # IFS=.-: 00:06:58.128 18:57:15 thread -- scripts/common.sh@336 -- # read -ra ver1 00:06:58.128 18:57:15 thread -- scripts/common.sh@337 -- # IFS=.-: 00:06:58.128 18:57:15 thread -- scripts/common.sh@337 -- # read -ra ver2 00:06:58.128 18:57:15 thread -- scripts/common.sh@338 -- # local 'op=<' 00:06:58.128 18:57:15 thread -- scripts/common.sh@340 -- # ver1_l=2 00:06:58.128 18:57:15 thread -- scripts/common.sh@341 -- # ver2_l=1 00:06:58.128 18:57:15 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:58.128 18:57:15 thread -- scripts/common.sh@344 -- # case "$op" in 00:06:58.128 18:57:15 thread -- scripts/common.sh@345 -- # : 1 00:06:58.128 18:57:15 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:58.128 18:57:15 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:58.128 18:57:15 thread -- scripts/common.sh@365 -- # decimal 1 00:06:58.128 18:57:15 thread -- scripts/common.sh@353 -- # local d=1 00:06:58.128 18:57:15 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:58.128 18:57:15 thread -- scripts/common.sh@355 -- # echo 1 00:06:58.128 18:57:15 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:06:58.128 18:57:15 thread -- scripts/common.sh@366 -- # decimal 2 00:06:58.128 18:57:15 thread -- scripts/common.sh@353 -- # local d=2 00:06:58.129 18:57:15 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:58.129 18:57:15 thread -- scripts/common.sh@355 -- # echo 2 00:06:58.129 18:57:15 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:06:58.129 18:57:15 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:58.129 18:57:15 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:58.129 18:57:15 thread -- scripts/common.sh@368 -- # return 0 00:06:58.129 18:57:15 thread -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:58.129 18:57:15 thread -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:06:58.129 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.129 --rc genhtml_branch_coverage=1 00:06:58.129 --rc genhtml_function_coverage=1 00:06:58.129 --rc genhtml_legend=1 00:06:58.129 --rc geninfo_all_blocks=1 00:06:58.129 --rc geninfo_unexecuted_blocks=1 00:06:58.129 00:06:58.129 ' 00:06:58.129 18:57:15 thread -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:06:58.129 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.129 --rc genhtml_branch_coverage=1 00:06:58.129 --rc genhtml_function_coverage=1 00:06:58.129 --rc genhtml_legend=1 00:06:58.129 --rc geninfo_all_blocks=1 00:06:58.129 --rc geninfo_unexecuted_blocks=1 00:06:58.129 00:06:58.129 ' 00:06:58.129 18:57:15 thread -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:06:58.129 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.129 --rc genhtml_branch_coverage=1 00:06:58.129 --rc genhtml_function_coverage=1 00:06:58.129 --rc genhtml_legend=1 00:06:58.129 --rc geninfo_all_blocks=1 00:06:58.129 --rc geninfo_unexecuted_blocks=1 00:06:58.129 00:06:58.129 ' 00:06:58.129 18:57:15 thread -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:06:58.129 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:58.129 --rc genhtml_branch_coverage=1 00:06:58.129 --rc genhtml_function_coverage=1 00:06:58.129 --rc genhtml_legend=1 00:06:58.129 --rc geninfo_all_blocks=1 00:06:58.129 --rc geninfo_unexecuted_blocks=1 00:06:58.129 00:06:58.129 ' 00:06:58.129 18:57:15 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:06:58.129 18:57:15 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:06:58.129 18:57:15 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:58.129 18:57:15 thread -- common/autotest_common.sh@10 -- # set +x 00:06:58.129 ************************************ 00:06:58.129 START TEST thread_poller_perf 00:06:58.129 ************************************ 00:06:58.129 18:57:15 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:06:58.389 [2024-12-05 18:57:15.719624] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:58.389 [2024-12-05 18:57:15.719830] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71012 ] 00:06:58.389 [2024-12-05 18:57:15.875753] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:58.389 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:06:58.389 [2024-12-05 18:57:15.903854] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:06:59.770 [2024-12-05T18:57:17.329Z] ====================================== 00:06:59.770 [2024-12-05T18:57:17.329Z] busy:2300469378 (cyc) 00:06:59.770 [2024-12-05T18:57:17.329Z] total_run_count: 430000 00:06:59.770 [2024-12-05T18:57:17.329Z] tsc_hz: 2290000000 (cyc) 00:06:59.770 [2024-12-05T18:57:17.329Z] ====================================== 00:06:59.770 [2024-12-05T18:57:17.329Z] poller_cost: 5349 (cyc), 2335 (nsec) 00:06:59.770 00:06:59.770 real 0m1.287s 00:06:59.770 user 0m1.115s 00:06:59.770 sys 0m0.067s 00:06:59.770 18:57:16 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:06:59.770 ************************************ 00:06:59.770 END TEST thread_poller_perf 00:06:59.770 ************************************ 00:06:59.770 18:57:16 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:06:59.770 18:57:17 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:06:59.770 18:57:17 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:06:59.770 18:57:17 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:06:59.770 18:57:17 thread -- common/autotest_common.sh@10 -- # set +x 00:06:59.770 ************************************ 00:06:59.770 START TEST thread_poller_perf 00:06:59.770 ************************************ 00:06:59.770 18:57:17 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:06:59.770 [2024-12-05 18:57:17.079895] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:06:59.770 [2024-12-05 18:57:17.080071] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71045 ] 00:06:59.770 [2024-12-05 18:57:17.234085] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:59.770 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:06:59.770 [2024-12-05 18:57:17.260264] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:01.151 [2024-12-05T18:57:18.710Z] ====================================== 00:07:01.151 [2024-12-05T18:57:18.710Z] busy:2293487126 (cyc) 00:07:01.151 [2024-12-05T18:57:18.710Z] total_run_count: 5663000 00:07:01.151 [2024-12-05T18:57:18.710Z] tsc_hz: 2290000000 (cyc) 00:07:01.151 [2024-12-05T18:57:18.710Z] ====================================== 00:07:01.151 [2024-12-05T18:57:18.710Z] poller_cost: 404 (cyc), 176 (nsec) 00:07:01.151 00:07:01.151 real 0m1.282s 00:07:01.151 user 0m1.096s 00:07:01.151 sys 0m0.081s 00:07:01.151 18:57:18 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:01.151 18:57:18 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:07:01.151 ************************************ 00:07:01.151 END TEST thread_poller_perf 00:07:01.151 ************************************ 00:07:01.151 18:57:18 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:07:01.151 ************************************ 00:07:01.151 END TEST thread 00:07:01.151 ************************************ 00:07:01.151 00:07:01.151 real 0m2.925s 00:07:01.151 user 0m2.383s 00:07:01.151 sys 0m0.341s 00:07:01.151 18:57:18 thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:01.151 18:57:18 thread -- common/autotest_common.sh@10 -- # set +x 00:07:01.151 18:57:18 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:07:01.151 18:57:18 -- spdk/autotest.sh@176 -- # run_test app_cmdline /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:07:01.151 18:57:18 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:01.151 18:57:18 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:01.151 18:57:18 -- common/autotest_common.sh@10 -- # set +x 00:07:01.151 ************************************ 00:07:01.151 START TEST app_cmdline 00:07:01.151 ************************************ 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:07:01.151 * Looking for test storage... 00:07:01.151 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@1711 -- # lcov --version 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@345 -- # : 1 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:01.151 18:57:18 app_cmdline -- scripts/common.sh@368 -- # return 0 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:07:01.151 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:01.151 --rc genhtml_branch_coverage=1 00:07:01.151 --rc genhtml_function_coverage=1 00:07:01.151 --rc genhtml_legend=1 00:07:01.151 --rc geninfo_all_blocks=1 00:07:01.151 --rc geninfo_unexecuted_blocks=1 00:07:01.151 00:07:01.151 ' 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:07:01.151 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:01.151 --rc genhtml_branch_coverage=1 00:07:01.151 --rc genhtml_function_coverage=1 00:07:01.151 --rc genhtml_legend=1 00:07:01.151 --rc geninfo_all_blocks=1 00:07:01.151 --rc geninfo_unexecuted_blocks=1 00:07:01.151 00:07:01.151 ' 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:07:01.151 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:01.151 --rc genhtml_branch_coverage=1 00:07:01.151 --rc genhtml_function_coverage=1 00:07:01.151 --rc genhtml_legend=1 00:07:01.151 --rc geninfo_all_blocks=1 00:07:01.151 --rc geninfo_unexecuted_blocks=1 00:07:01.151 00:07:01.151 ' 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:07:01.151 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:01.151 --rc genhtml_branch_coverage=1 00:07:01.151 --rc genhtml_function_coverage=1 00:07:01.151 --rc genhtml_legend=1 00:07:01.151 --rc geninfo_all_blocks=1 00:07:01.151 --rc geninfo_unexecuted_blocks=1 00:07:01.151 00:07:01.151 ' 00:07:01.151 18:57:18 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:07:01.151 18:57:18 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=71133 00:07:01.151 18:57:18 app_cmdline -- app/cmdline.sh@16 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:07:01.151 18:57:18 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 71133 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@835 -- # '[' -z 71133 ']' 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:01.151 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:01.151 18:57:18 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:07:01.412 [2024-12-05 18:57:18.756796] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:01.412 [2024-12-05 18:57:18.757005] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71133 ] 00:07:01.412 [2024-12-05 18:57:18.910836] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:01.412 [2024-12-05 18:57:18.935626] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@868 -- # return 0 00:07:02.350 18:57:19 app_cmdline -- app/cmdline.sh@20 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py spdk_get_version 00:07:02.350 { 00:07:02.350 "version": "SPDK v25.01-pre git sha1 8d3947977", 00:07:02.350 "fields": { 00:07:02.350 "major": 25, 00:07:02.350 "minor": 1, 00:07:02.350 "patch": 0, 00:07:02.350 "suffix": "-pre", 00:07:02.350 "commit": "8d3947977" 00:07:02.350 } 00:07:02.350 } 00:07:02.350 18:57:19 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:07:02.350 18:57:19 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:07:02.350 18:57:19 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:07:02.350 18:57:19 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:07:02.350 18:57:19 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:07:02.350 18:57:19 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:07:02.350 18:57:19 app_cmdline -- app/cmdline.sh@26 -- # sort 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:02.350 18:57:19 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:07:02.350 18:57:19 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:07:02.350 18:57:19 app_cmdline -- app/cmdline.sh@30 -- # NOT /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@652 -- # local es=0 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@654 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@640 -- # local arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@644 -- # type -t /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@646 -- # type -P /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@646 -- # arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@646 -- # [[ -x /home/vagrant/spdk_repo/spdk/scripts/rpc.py ]] 00:07:02.350 18:57:19 app_cmdline -- common/autotest_common.sh@655 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:07:02.609 request: 00:07:02.609 { 00:07:02.609 "method": "env_dpdk_get_mem_stats", 00:07:02.609 "req_id": 1 00:07:02.609 } 00:07:02.609 Got JSON-RPC error response 00:07:02.609 response: 00:07:02.609 { 00:07:02.609 "code": -32601, 00:07:02.609 "message": "Method not found" 00:07:02.609 } 00:07:02.609 18:57:19 app_cmdline -- common/autotest_common.sh@655 -- # es=1 00:07:02.609 18:57:19 app_cmdline -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:02.609 18:57:19 app_cmdline -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:02.609 18:57:19 app_cmdline -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:02.609 18:57:19 app_cmdline -- app/cmdline.sh@1 -- # killprocess 71133 00:07:02.609 18:57:19 app_cmdline -- common/autotest_common.sh@954 -- # '[' -z 71133 ']' 00:07:02.609 18:57:19 app_cmdline -- common/autotest_common.sh@958 -- # kill -0 71133 00:07:02.609 18:57:19 app_cmdline -- common/autotest_common.sh@959 -- # uname 00:07:02.609 18:57:19 app_cmdline -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:02.609 18:57:19 app_cmdline -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71133 00:07:02.609 killing process with pid 71133 00:07:02.609 18:57:20 app_cmdline -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:02.609 18:57:20 app_cmdline -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:02.609 18:57:20 app_cmdline -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71133' 00:07:02.609 18:57:20 app_cmdline -- common/autotest_common.sh@973 -- # kill 71133 00:07:02.609 18:57:20 app_cmdline -- common/autotest_common.sh@978 -- # wait 71133 00:07:02.868 ************************************ 00:07:02.868 END TEST app_cmdline 00:07:02.868 ************************************ 00:07:02.868 00:07:02.868 real 0m1.926s 00:07:02.868 user 0m2.130s 00:07:02.868 sys 0m0.528s 00:07:02.868 18:57:20 app_cmdline -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:02.868 18:57:20 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:07:03.127 18:57:20 -- spdk/autotest.sh@177 -- # run_test version /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:07:03.127 18:57:20 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:03.127 18:57:20 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:03.127 18:57:20 -- common/autotest_common.sh@10 -- # set +x 00:07:03.127 ************************************ 00:07:03.127 START TEST version 00:07:03.127 ************************************ 00:07:03.127 18:57:20 version -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:07:03.127 * Looking for test storage... 00:07:03.127 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:07:03.127 18:57:20 version -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:07:03.127 18:57:20 version -- common/autotest_common.sh@1711 -- # lcov --version 00:07:03.127 18:57:20 version -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:07:03.127 18:57:20 version -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:07:03.127 18:57:20 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:03.127 18:57:20 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:03.127 18:57:20 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:03.127 18:57:20 version -- scripts/common.sh@336 -- # IFS=.-: 00:07:03.127 18:57:20 version -- scripts/common.sh@336 -- # read -ra ver1 00:07:03.127 18:57:20 version -- scripts/common.sh@337 -- # IFS=.-: 00:07:03.127 18:57:20 version -- scripts/common.sh@337 -- # read -ra ver2 00:07:03.127 18:57:20 version -- scripts/common.sh@338 -- # local 'op=<' 00:07:03.127 18:57:20 version -- scripts/common.sh@340 -- # ver1_l=2 00:07:03.127 18:57:20 version -- scripts/common.sh@341 -- # ver2_l=1 00:07:03.127 18:57:20 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:03.127 18:57:20 version -- scripts/common.sh@344 -- # case "$op" in 00:07:03.127 18:57:20 version -- scripts/common.sh@345 -- # : 1 00:07:03.127 18:57:20 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:03.127 18:57:20 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:03.127 18:57:20 version -- scripts/common.sh@365 -- # decimal 1 00:07:03.127 18:57:20 version -- scripts/common.sh@353 -- # local d=1 00:07:03.127 18:57:20 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:03.127 18:57:20 version -- scripts/common.sh@355 -- # echo 1 00:07:03.127 18:57:20 version -- scripts/common.sh@365 -- # ver1[v]=1 00:07:03.127 18:57:20 version -- scripts/common.sh@366 -- # decimal 2 00:07:03.127 18:57:20 version -- scripts/common.sh@353 -- # local d=2 00:07:03.127 18:57:20 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:03.127 18:57:20 version -- scripts/common.sh@355 -- # echo 2 00:07:03.127 18:57:20 version -- scripts/common.sh@366 -- # ver2[v]=2 00:07:03.127 18:57:20 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:03.127 18:57:20 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:03.127 18:57:20 version -- scripts/common.sh@368 -- # return 0 00:07:03.127 18:57:20 version -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:03.127 18:57:20 version -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:07:03.127 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:03.127 --rc genhtml_branch_coverage=1 00:07:03.127 --rc genhtml_function_coverage=1 00:07:03.127 --rc genhtml_legend=1 00:07:03.127 --rc geninfo_all_blocks=1 00:07:03.127 --rc geninfo_unexecuted_blocks=1 00:07:03.127 00:07:03.127 ' 00:07:03.127 18:57:20 version -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:07:03.127 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:03.127 --rc genhtml_branch_coverage=1 00:07:03.127 --rc genhtml_function_coverage=1 00:07:03.127 --rc genhtml_legend=1 00:07:03.127 --rc geninfo_all_blocks=1 00:07:03.127 --rc geninfo_unexecuted_blocks=1 00:07:03.127 00:07:03.127 ' 00:07:03.127 18:57:20 version -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:07:03.127 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:03.127 --rc genhtml_branch_coverage=1 00:07:03.127 --rc genhtml_function_coverage=1 00:07:03.127 --rc genhtml_legend=1 00:07:03.127 --rc geninfo_all_blocks=1 00:07:03.127 --rc geninfo_unexecuted_blocks=1 00:07:03.127 00:07:03.127 ' 00:07:03.127 18:57:20 version -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:07:03.127 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:03.127 --rc genhtml_branch_coverage=1 00:07:03.127 --rc genhtml_function_coverage=1 00:07:03.127 --rc genhtml_legend=1 00:07:03.127 --rc geninfo_all_blocks=1 00:07:03.127 --rc geninfo_unexecuted_blocks=1 00:07:03.127 00:07:03.127 ' 00:07:03.127 18:57:20 version -- app/version.sh@17 -- # get_header_version major 00:07:03.127 18:57:20 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:03.127 18:57:20 version -- app/version.sh@14 -- # cut -f2 00:07:03.127 18:57:20 version -- app/version.sh@14 -- # tr -d '"' 00:07:03.127 18:57:20 version -- app/version.sh@17 -- # major=25 00:07:03.127 18:57:20 version -- app/version.sh@18 -- # get_header_version minor 00:07:03.127 18:57:20 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:03.127 18:57:20 version -- app/version.sh@14 -- # cut -f2 00:07:03.127 18:57:20 version -- app/version.sh@14 -- # tr -d '"' 00:07:03.127 18:57:20 version -- app/version.sh@18 -- # minor=1 00:07:03.127 18:57:20 version -- app/version.sh@19 -- # get_header_version patch 00:07:03.127 18:57:20 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:03.127 18:57:20 version -- app/version.sh@14 -- # cut -f2 00:07:03.386 18:57:20 version -- app/version.sh@14 -- # tr -d '"' 00:07:03.386 18:57:20 version -- app/version.sh@19 -- # patch=0 00:07:03.386 18:57:20 version -- app/version.sh@20 -- # get_header_version suffix 00:07:03.386 18:57:20 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:03.386 18:57:20 version -- app/version.sh@14 -- # cut -f2 00:07:03.387 18:57:20 version -- app/version.sh@14 -- # tr -d '"' 00:07:03.387 18:57:20 version -- app/version.sh@20 -- # suffix=-pre 00:07:03.387 18:57:20 version -- app/version.sh@22 -- # version=25.1 00:07:03.387 18:57:20 version -- app/version.sh@25 -- # (( patch != 0 )) 00:07:03.387 18:57:20 version -- app/version.sh@28 -- # version=25.1rc0 00:07:03.387 18:57:20 version -- app/version.sh@30 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python 00:07:03.387 18:57:20 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:07:03.387 18:57:20 version -- app/version.sh@30 -- # py_version=25.1rc0 00:07:03.387 18:57:20 version -- app/version.sh@31 -- # [[ 25.1rc0 == \2\5\.\1\r\c\0 ]] 00:07:03.387 ************************************ 00:07:03.387 END TEST version 00:07:03.387 ************************************ 00:07:03.387 00:07:03.387 real 0m0.313s 00:07:03.387 user 0m0.181s 00:07:03.387 sys 0m0.188s 00:07:03.387 18:57:20 version -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:03.387 18:57:20 version -- common/autotest_common.sh@10 -- # set +x 00:07:03.387 18:57:20 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:07:03.387 18:57:20 -- spdk/autotest.sh@188 -- # [[ 1 -eq 1 ]] 00:07:03.387 18:57:20 -- spdk/autotest.sh@189 -- # run_test bdev_raid /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:07:03.387 18:57:20 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:03.387 18:57:20 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:03.387 18:57:20 -- common/autotest_common.sh@10 -- # set +x 00:07:03.387 ************************************ 00:07:03.387 START TEST bdev_raid 00:07:03.387 ************************************ 00:07:03.387 18:57:20 bdev_raid -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:07:03.387 * Looking for test storage... 00:07:03.387 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:07:03.387 18:57:20 bdev_raid -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:07:03.387 18:57:20 bdev_raid -- common/autotest_common.sh@1711 -- # lcov --version 00:07:03.387 18:57:20 bdev_raid -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:07:03.644 18:57:20 bdev_raid -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:07:03.644 18:57:20 bdev_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:03.644 18:57:20 bdev_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:03.644 18:57:20 bdev_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:03.644 18:57:20 bdev_raid -- scripts/common.sh@336 -- # IFS=.-: 00:07:03.644 18:57:20 bdev_raid -- scripts/common.sh@336 -- # read -ra ver1 00:07:03.644 18:57:20 bdev_raid -- scripts/common.sh@337 -- # IFS=.-: 00:07:03.645 18:57:20 bdev_raid -- scripts/common.sh@337 -- # read -ra ver2 00:07:03.645 18:57:20 bdev_raid -- scripts/common.sh@338 -- # local 'op=<' 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@340 -- # ver1_l=2 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@341 -- # ver2_l=1 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@344 -- # case "$op" in 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@345 -- # : 1 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@365 -- # decimal 1 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@353 -- # local d=1 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@355 -- # echo 1 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@366 -- # decimal 2 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@353 -- # local d=2 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@355 -- # echo 2 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:03.645 18:57:21 bdev_raid -- scripts/common.sh@368 -- # return 0 00:07:03.645 18:57:21 bdev_raid -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:03.645 18:57:21 bdev_raid -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:07:03.645 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:03.645 --rc genhtml_branch_coverage=1 00:07:03.645 --rc genhtml_function_coverage=1 00:07:03.645 --rc genhtml_legend=1 00:07:03.645 --rc geninfo_all_blocks=1 00:07:03.645 --rc geninfo_unexecuted_blocks=1 00:07:03.645 00:07:03.645 ' 00:07:03.645 18:57:21 bdev_raid -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:07:03.645 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:03.645 --rc genhtml_branch_coverage=1 00:07:03.645 --rc genhtml_function_coverage=1 00:07:03.645 --rc genhtml_legend=1 00:07:03.645 --rc geninfo_all_blocks=1 00:07:03.645 --rc geninfo_unexecuted_blocks=1 00:07:03.645 00:07:03.645 ' 00:07:03.645 18:57:21 bdev_raid -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:07:03.645 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:03.645 --rc genhtml_branch_coverage=1 00:07:03.645 --rc genhtml_function_coverage=1 00:07:03.645 --rc genhtml_legend=1 00:07:03.645 --rc geninfo_all_blocks=1 00:07:03.645 --rc geninfo_unexecuted_blocks=1 00:07:03.645 00:07:03.645 ' 00:07:03.645 18:57:21 bdev_raid -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:07:03.645 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:03.645 --rc genhtml_branch_coverage=1 00:07:03.645 --rc genhtml_function_coverage=1 00:07:03.645 --rc genhtml_legend=1 00:07:03.645 --rc geninfo_all_blocks=1 00:07:03.645 --rc geninfo_unexecuted_blocks=1 00:07:03.645 00:07:03.645 ' 00:07:03.645 18:57:21 bdev_raid -- bdev/bdev_raid.sh@12 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:07:03.645 18:57:21 bdev_raid -- bdev/nbd_common.sh@6 -- # set -e 00:07:03.645 18:57:21 bdev_raid -- bdev/bdev_raid.sh@14 -- # rpc_py=rpc_cmd 00:07:03.645 18:57:21 bdev_raid -- bdev/bdev_raid.sh@946 -- # mkdir -p /raidtest 00:07:03.645 18:57:21 bdev_raid -- bdev/bdev_raid.sh@947 -- # trap 'cleanup; exit 1' EXIT 00:07:03.645 18:57:21 bdev_raid -- bdev/bdev_raid.sh@949 -- # base_blocklen=512 00:07:03.645 18:57:21 bdev_raid -- bdev/bdev_raid.sh@951 -- # run_test raid1_resize_data_offset_test raid_resize_data_offset_test 00:07:03.645 18:57:21 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:07:03.645 18:57:21 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:03.645 18:57:21 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:03.645 ************************************ 00:07:03.645 START TEST raid1_resize_data_offset_test 00:07:03.645 ************************************ 00:07:03.645 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1129 -- # raid_resize_data_offset_test 00:07:03.645 18:57:21 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@917 -- # raid_pid=71294 00:07:03.645 18:57:21 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@918 -- # echo 'Process raid pid: 71294' 00:07:03.645 Process raid pid: 71294 00:07:03.645 18:57:21 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@916 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:03.645 18:57:21 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@919 -- # waitforlisten 71294 00:07:03.645 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@835 -- # '[' -z 71294 ']' 00:07:03.645 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:03.645 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:03.645 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:03.645 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:03.645 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:03.645 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:03.645 [2024-12-05 18:57:21.127586] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:03.645 [2024-12-05 18:57:21.127814] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:03.904 [2024-12-05 18:57:21.283546] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:03.904 [2024-12-05 18:57:21.308038] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:03.904 [2024-12-05 18:57:21.349863] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:03.904 [2024-12-05 18:57:21.349974] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@868 -- # return 0 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@922 -- # rpc_cmd bdev_malloc_create -b malloc0 64 512 -o 16 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.473 malloc0 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@923 -- # rpc_cmd bdev_malloc_create -b malloc1 64 512 -o 16 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.473 malloc1 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@924 -- # rpc_cmd bdev_null_create null0 64 512 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.473 null0 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@926 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''malloc0 malloc1 null0'\''' -s 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.473 [2024-12-05 18:57:21.994890] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc0 is claimed 00:07:04.473 [2024-12-05 18:57:21.996808] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:04.473 [2024-12-05 18:57:21.996891] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev null0 is claimed 00:07:04.473 [2024-12-05 18:57:21.997052] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:04.473 [2024-12-05 18:57:21.997097] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 129024, blocklen 512 00:07:04.473 [2024-12-05 18:57:21.997373] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:07:04.473 [2024-12-05 18:57:21.997516] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:04.473 [2024-12-05 18:57:21.997568] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:07:04.473 [2024-12-05 18:57:21.997728] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:04.473 18:57:21 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.473 18:57:22 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:07:04.473 18:57:22 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:04.473 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.473 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.473 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # (( 2048 == 2048 )) 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@931 -- # rpc_cmd bdev_null_delete null0 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.733 [2024-12-05 18:57:22.058791] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: null0 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@935 -- # rpc_cmd bdev_malloc_create -b malloc2 512 512 -o 30 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.733 malloc2 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@936 -- # rpc_cmd bdev_raid_add_base_bdev Raid malloc2 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.733 [2024-12-05 18:57:22.186595] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:04.733 [2024-12-05 18:57:22.193037] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.733 [2024-12-05 18:57:22.196226] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev Raid 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # (( 2070 == 2070 )) 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@941 -- # killprocess 71294 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@954 -- # '[' -z 71294 ']' 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@958 -- # kill -0 71294 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@959 -- # uname 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71294 00:07:04.733 killing process with pid 71294 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71294' 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@973 -- # kill 71294 00:07:04.733 [2024-12-05 18:57:22.286460] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:04.733 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@978 -- # wait 71294 00:07:04.733 [2024-12-05 18:57:22.287455] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev Raid: Operation canceled 00:07:04.733 [2024-12-05 18:57:22.287528] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:04.733 [2024-12-05 18:57:22.287562] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: malloc2 00:07:04.993 [2024-12-05 18:57:22.293393] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:04.993 [2024-12-05 18:57:22.293748] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:04.993 [2024-12-05 18:57:22.293770] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:07:04.993 [2024-12-05 18:57:22.498770] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:05.252 ************************************ 00:07:05.252 END TEST raid1_resize_data_offset_test 00:07:05.252 ************************************ 00:07:05.252 18:57:22 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@943 -- # return 0 00:07:05.252 00:07:05.252 real 0m1.661s 00:07:05.252 user 0m1.647s 00:07:05.252 sys 0m0.433s 00:07:05.252 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:05.252 18:57:22 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:05.252 18:57:22 bdev_raid -- bdev/bdev_raid.sh@953 -- # run_test raid0_resize_superblock_test raid_resize_superblock_test 0 00:07:05.252 18:57:22 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:07:05.252 18:57:22 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:05.252 18:57:22 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:05.252 ************************************ 00:07:05.252 START TEST raid0_resize_superblock_test 00:07:05.252 ************************************ 00:07:05.252 18:57:22 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1129 -- # raid_resize_superblock_test 0 00:07:05.252 18:57:22 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=0 00:07:05.252 18:57:22 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71345 00:07:05.252 18:57:22 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:05.252 18:57:22 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71345' 00:07:05.252 Process raid pid: 71345 00:07:05.252 18:57:22 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71345 00:07:05.252 18:57:22 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 71345 ']' 00:07:05.252 18:57:22 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:05.252 18:57:22 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:05.252 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:05.252 18:57:22 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:05.252 18:57:22 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:05.252 18:57:22 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:05.512 [2024-12-05 18:57:22.861958] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:05.512 [2024-12-05 18:57:22.862126] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:05.512 [2024-12-05 18:57:23.019241] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:05.512 [2024-12-05 18:57:23.043502] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:05.772 [2024-12-05 18:57:23.086419] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:05.772 [2024-12-05 18:57:23.086544] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:06.342 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:06.342 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:07:06.342 18:57:23 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:07:06.342 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.342 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.342 malloc0 00:07:06.342 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.342 18:57:23 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:06.342 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.342 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.342 [2024-12-05 18:57:23.805567] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:06.342 [2024-12-05 18:57:23.805632] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:06.342 [2024-12-05 18:57:23.805674] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:06.342 [2024-12-05 18:57:23.805703] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:06.342 [2024-12-05 18:57:23.807890] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:06.342 [2024-12-05 18:57:23.807971] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:06.342 pt0 00:07:06.342 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.342 18:57:23 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:07:06.342 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.342 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.603 fd8f6e75-cbeb-4ce5-ad62-9e53649e99ec 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.603 47256abb-2e83-4c98-9cb4-efb9afbc33a1 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.603 b77a028a-82fc-4932-85c8-b776952b46cf 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@870 -- # rpc_cmd bdev_raid_create -n Raid -r 0 -z 64 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.603 [2024-12-05 18:57:23.939629] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 47256abb-2e83-4c98-9cb4-efb9afbc33a1 is claimed 00:07:06.603 [2024-12-05 18:57:23.939746] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev b77a028a-82fc-4932-85c8-b776952b46cf is claimed 00:07:06.603 [2024-12-05 18:57:23.939859] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:06.603 [2024-12-05 18:57:23.939873] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 245760, blocklen 512 00:07:06.603 [2024-12-05 18:57:23.940142] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:06.603 [2024-12-05 18:57:23.940297] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:06.603 [2024-12-05 18:57:23.940307] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:07:06.603 [2024-12-05 18:57:23.940429] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.603 18:57:23 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # jq '.[].num_blocks' 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.603 [2024-12-05 18:57:24.051688] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # (( 245760 == 245760 )) 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.603 [2024-12-05 18:57:24.099498] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:06.603 [2024-12-05 18:57:24.099575] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '47256abb-2e83-4c98-9cb4-efb9afbc33a1' was resized: old size 131072, new size 204800 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.603 [2024-12-05 18:57:24.111431] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:06.603 [2024-12-05 18:57:24.111501] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'b77a028a-82fc-4932-85c8-b776952b46cf' was resized: old size 131072, new size 204800 00:07:06.603 [2024-12-05 18:57:24.111569] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 245760 to 393216 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.603 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # jq '.[].num_blocks' 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.864 [2024-12-05 18:57:24.215352] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # (( 393216 == 393216 )) 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.864 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.864 [2024-12-05 18:57:24.263072] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:07:06.864 [2024-12-05 18:57:24.263181] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:07:06.864 [2024-12-05 18:57:24.263196] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:06.864 [2024-12-05 18:57:24.263215] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:07:06.865 [2024-12-05 18:57:24.263326] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:06.865 [2024-12-05 18:57:24.263373] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:06.865 [2024-12-05 18:57:24.263387] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.865 [2024-12-05 18:57:24.275018] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:06.865 [2024-12-05 18:57:24.275085] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:06.865 [2024-12-05 18:57:24.275103] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:06.865 [2024-12-05 18:57:24.275114] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:06.865 [2024-12-05 18:57:24.277177] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:06.865 [2024-12-05 18:57:24.277214] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:06.865 [2024-12-05 18:57:24.278545] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 47256abb-2e83-4c98-9cb4-efb9afbc33a1 00:07:06.865 [2024-12-05 18:57:24.278647] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 47256abb-2e83-4c98-9cb4-efb9afbc33a1 is claimed 00:07:06.865 [2024-12-05 18:57:24.278751] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev b77a028a-82fc-4932-85c8-b776952b46cf 00:07:06.865 [2024-12-05 18:57:24.278784] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev b77a028a-82fc-4932-85c8-b776952b46cf is claimed 00:07:06.865 [2024-12-05 18:57:24.278871] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev b77a028a-82fc-4932-85c8-b776952b46cf (2) smaller than existing raid bdev Raid (3) 00:07:06.865 [2024-12-05 18:57:24.278895] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 47256abb-2e83-4c98-9cb4-efb9afbc33a1: File exists 00:07:06.865 [2024-12-05 18:57:24.278931] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:07:06.865 [2024-12-05 18:57:24.278940] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 393216, blocklen 512 00:07:06.865 [2024-12-05 18:57:24.279154] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:07:06.865 [2024-12-05 18:57:24.279312] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:07:06.865 [2024-12-05 18:57:24.279326] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001580 00:07:06.865 [2024-12-05 18:57:24.279476] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:06.865 pt0 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # jq '.[].num_blocks' 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.865 [2024-12-05 18:57:24.299613] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # (( 393216 == 393216 )) 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71345 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 71345 ']' 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@958 -- # kill -0 71345 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@959 -- # uname 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71345 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71345' 00:07:06.865 killing process with pid 71345 00:07:06.865 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@973 -- # kill 71345 00:07:06.865 [2024-12-05 18:57:24.376946] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:06.865 [2024-12-05 18:57:24.377043] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:06.865 [2024-12-05 18:57:24.377105] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:06.865 [2024-12-05 18:57:24.377149] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Raid, state off 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@978 -- # wait 71345 00:07:06.865 line 00:07:07.177 [2024-12-05 18:57:24.534345] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:07.455 18:57:24 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:07:07.455 00:07:07.455 real 0m1.968s 00:07:07.455 user 0m2.245s 00:07:07.455 sys 0m0.477s 00:07:07.455 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:07.455 18:57:24 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.455 ************************************ 00:07:07.455 END TEST raid0_resize_superblock_test 00:07:07.455 ************************************ 00:07:07.455 18:57:24 bdev_raid -- bdev/bdev_raid.sh@954 -- # run_test raid1_resize_superblock_test raid_resize_superblock_test 1 00:07:07.455 18:57:24 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:07:07.455 18:57:24 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:07.455 18:57:24 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:07.455 ************************************ 00:07:07.455 START TEST raid1_resize_superblock_test 00:07:07.455 ************************************ 00:07:07.455 18:57:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1129 -- # raid_resize_superblock_test 1 00:07:07.455 18:57:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=1 00:07:07.455 18:57:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71416 00:07:07.455 18:57:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:07.455 18:57:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71416' 00:07:07.455 Process raid pid: 71416 00:07:07.455 18:57:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71416 00:07:07.455 18:57:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 71416 ']' 00:07:07.455 18:57:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:07.455 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:07.455 18:57:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:07.455 18:57:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:07.455 18:57:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:07.455 18:57:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.455 [2024-12-05 18:57:24.897308] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:07.455 [2024-12-05 18:57:24.897444] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:07.716 [2024-12-05 18:57:25.051282] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:07.716 [2024-12-05 18:57:25.075438] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:07.716 [2024-12-05 18:57:25.117063] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:07.716 [2024-12-05 18:57:25.117184] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:08.284 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:08.284 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:07:08.284 18:57:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:07:08.284 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.284 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.543 malloc0 00:07:08.543 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.543 18:57:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:08.543 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.543 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.543 [2024-12-05 18:57:25.848026] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:08.544 [2024-12-05 18:57:25.848084] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:08.544 [2024-12-05 18:57:25.848123] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:08.544 [2024-12-05 18:57:25.848133] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:08.544 [2024-12-05 18:57:25.850231] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:08.544 [2024-12-05 18:57:25.850329] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:08.544 pt0 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.544 28e9c24b-eb63-4d81-a70d-5901cf8e6037 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.544 5d4b0d90-bbcf-464a-a908-47590d298ec4 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.544 e00448fd-8684-4856-8b61-ebbc449f5350 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@871 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.544 [2024-12-05 18:57:25.982067] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 5d4b0d90-bbcf-464a-a908-47590d298ec4 is claimed 00:07:08.544 [2024-12-05 18:57:25.982147] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev e00448fd-8684-4856-8b61-ebbc449f5350 is claimed 00:07:08.544 [2024-12-05 18:57:25.982249] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:08.544 [2024-12-05 18:57:25.982271] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 122880, blocklen 512 00:07:08.544 [2024-12-05 18:57:25.982511] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:08.544 [2024-12-05 18:57:25.982652] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:08.544 [2024-12-05 18:57:25.982681] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:07:08.544 [2024-12-05 18:57:25.982828] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.544 18:57:25 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.544 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.544 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:07:08.544 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:08.544 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:07:08.544 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.544 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.544 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.544 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:07:08.544 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:08.544 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # jq '.[].num_blocks' 00:07:08.544 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:08.544 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:08.544 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.544 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.544 [2024-12-05 18:57:26.094099] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:08.803 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.803 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:08.803 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:08.803 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # (( 122880 == 122880 )) 00:07:08.803 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:07:08.803 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.803 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.803 [2024-12-05 18:57:26.141913] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:08.803 [2024-12-05 18:57:26.141980] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '5d4b0d90-bbcf-464a-a908-47590d298ec4' was resized: old size 131072, new size 204800 00:07:08.803 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.803 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:07:08.803 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.803 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.804 [2024-12-05 18:57:26.153867] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:08.804 [2024-12-05 18:57:26.153889] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'e00448fd-8684-4856-8b61-ebbc449f5350' was resized: old size 131072, new size 204800 00:07:08.804 [2024-12-05 18:57:26.153916] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 122880 to 196608 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # jq '.[].num_blocks' 00:07:08.804 [2024-12-05 18:57:26.265802] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # (( 196608 == 196608 )) 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.804 [2024-12-05 18:57:26.313482] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:07:08.804 [2024-12-05 18:57:26.313553] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:07:08.804 [2024-12-05 18:57:26.313583] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:07:08.804 [2024-12-05 18:57:26.313729] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:08.804 [2024-12-05 18:57:26.313864] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:08.804 [2024-12-05 18:57:26.313917] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:08.804 [2024-12-05 18:57:26.313930] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.804 [2024-12-05 18:57:26.325439] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:08.804 [2024-12-05 18:57:26.325486] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:08.804 [2024-12-05 18:57:26.325518] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:08.804 [2024-12-05 18:57:26.325528] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:08.804 [2024-12-05 18:57:26.327652] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:08.804 [2024-12-05 18:57:26.327698] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:08.804 [2024-12-05 18:57:26.329028] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 5d4b0d90-bbcf-464a-a908-47590d298ec4 00:07:08.804 [2024-12-05 18:57:26.329076] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev 5d4b0d90-bbcf-464a-a908-47590d298ec4 is claimed 00:07:08.804 [2024-12-05 18:57:26.329143] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev e00448fd-8684-4856-8b61-ebbc449f5350 00:07:08.804 [2024-12-05 18:57:26.329171] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev e00448fd-8684-4856-8b61-ebbc449f5350 is claimed 00:07:08.804 [2024-12-05 18:57:26.329257] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev e00448fd-8684-4856-8b61-ebbc449f5350 (2) smaller than existing raid bdev Raid (3) 00:07:08.804 [2024-12-05 18:57:26.329275] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 5d4b0d90-bbcf-464a-a908-47590d298ec4: File exists 00:07:08.804 [2024-12-05 18:57:26.329316] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:07:08.804 [2024-12-05 18:57:26.329325] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:07:08.804 [2024-12-05 18:57:26.329529] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:07:08.804 [2024-12-05 18:57:26.329688] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:07:08.804 [2024-12-05 18:57:26.329703] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001580 00:07:08.804 [2024-12-05 18:57:26.329837] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:08.804 pt0 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # jq '.[].num_blocks' 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:08.804 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.804 [2024-12-05 18:57:26.353972] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:09.063 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:09.063 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:09.063 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:09.063 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # (( 196608 == 196608 )) 00:07:09.063 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71416 00:07:09.063 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 71416 ']' 00:07:09.063 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@958 -- # kill -0 71416 00:07:09.063 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@959 -- # uname 00:07:09.063 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:09.063 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71416 00:07:09.063 killing process with pid 71416 00:07:09.063 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:09.063 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:09.063 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71416' 00:07:09.063 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@973 -- # kill 71416 00:07:09.063 [2024-12-05 18:57:26.435989] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:09.063 [2024-12-05 18:57:26.436046] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:09.063 [2024-12-05 18:57:26.436087] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:09.063 [2024-12-05 18:57:26.436095] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Raid, state offline 00:07:09.063 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@978 -- # wait 71416 00:07:09.063 [2024-12-05 18:57:26.591957] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:09.321 18:57:26 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:07:09.321 00:07:09.322 real 0m1.984s 00:07:09.322 user 0m2.257s 00:07:09.322 sys 0m0.494s 00:07:09.322 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:09.322 ************************************ 00:07:09.322 END TEST raid1_resize_superblock_test 00:07:09.322 ************************************ 00:07:09.322 18:57:26 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.322 18:57:26 bdev_raid -- bdev/bdev_raid.sh@956 -- # uname -s 00:07:09.322 18:57:26 bdev_raid -- bdev/bdev_raid.sh@956 -- # '[' Linux = Linux ']' 00:07:09.322 18:57:26 bdev_raid -- bdev/bdev_raid.sh@956 -- # modprobe -n nbd 00:07:09.322 18:57:26 bdev_raid -- bdev/bdev_raid.sh@957 -- # has_nbd=true 00:07:09.322 18:57:26 bdev_raid -- bdev/bdev_raid.sh@958 -- # modprobe nbd 00:07:09.581 18:57:26 bdev_raid -- bdev/bdev_raid.sh@959 -- # run_test raid_function_test_raid0 raid_function_test raid0 00:07:09.581 18:57:26 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:07:09.581 18:57:26 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:09.581 18:57:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:09.581 ************************************ 00:07:09.581 START TEST raid_function_test_raid0 00:07:09.581 ************************************ 00:07:09.581 18:57:26 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1129 -- # raid_function_test raid0 00:07:09.581 18:57:26 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@64 -- # local raid_level=raid0 00:07:09.581 18:57:26 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:07:09.581 18:57:26 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:07:09.581 Process raid pid: 71491 00:07:09.581 18:57:26 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@69 -- # raid_pid=71491 00:07:09.581 18:57:26 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:09.581 18:57:26 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 71491' 00:07:09.581 18:57:26 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@71 -- # waitforlisten 71491 00:07:09.581 18:57:26 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@835 -- # '[' -z 71491 ']' 00:07:09.581 18:57:26 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:09.581 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:09.581 18:57:26 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:09.581 18:57:26 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:09.581 18:57:26 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:09.581 18:57:26 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:09.581 [2024-12-05 18:57:26.973654] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:09.581 [2024-12-05 18:57:26.973873] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:09.581 [2024-12-05 18:57:27.127633] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:09.840 [2024-12-05 18:57:27.152129] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:09.840 [2024-12-05 18:57:27.194029] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:09.840 [2024-12-05 18:57:27.194141] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:10.409 18:57:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:10.409 18:57:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@868 -- # return 0 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:10.410 Base_1 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:10.410 Base_2 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''Base_1 Base_2'\''' -n raid 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:10.410 [2024-12-05 18:57:27.821883] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:10.410 [2024-12-05 18:57:27.823810] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:10.410 [2024-12-05 18:57:27.823867] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:10.410 [2024-12-05 18:57:27.823879] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:10.410 [2024-12-05 18:57:27.824132] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:10.410 [2024-12-05 18:57:27.824245] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:10.410 [2024-12-05 18:57:27.824254] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000001200 00:07:10.410 [2024-12-05 18:57:27.824378] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@12 -- # local i 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:10.410 18:57:27 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:07:10.670 [2024-12-05 18:57:28.045569] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:10.670 /dev/nbd0 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@873 -- # local i 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@877 -- # break 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:07:10.670 1+0 records in 00:07:10.670 1+0 records out 00:07:10.670 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00041752 s, 9.8 MB/s 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@890 -- # size=4096 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@893 -- # return 0 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:10.670 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:10.930 { 00:07:10.930 "nbd_device": "/dev/nbd0", 00:07:10.930 "bdev_name": "raid" 00:07:10.930 } 00:07:10.930 ]' 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:10.930 { 00:07:10.930 "nbd_device": "/dev/nbd0", 00:07:10.930 "bdev_name": "raid" 00:07:10.930 } 00:07:10.930 ]' 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=1 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 1 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # count=1 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@19 -- # local blksize 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # blksize=512 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:07:10.930 4096+0 records in 00:07:10.930 4096+0 records out 00:07:10.930 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0347479 s, 60.4 MB/s 00:07:10.930 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:07:11.191 4096+0 records in 00:07:11.191 4096+0 records out 00:07:11.191 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.179671 s, 11.7 MB/s 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:07:11.191 128+0 records in 00:07:11.191 128+0 records out 00:07:11.191 65536 bytes (66 kB, 64 KiB) copied, 0.00115161 s, 56.9 MB/s 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:07:11.191 2035+0 records in 00:07:11.191 2035+0 records out 00:07:11.191 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0157486 s, 66.2 MB/s 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:07:11.191 456+0 records in 00:07:11.191 456+0 records out 00:07:11.191 233472 bytes (233 kB, 228 KiB) copied, 0.00272512 s, 85.7 MB/s 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@52 -- # return 0 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:07:11.191 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:11.192 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@51 -- # local i 00:07:11.192 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:11.192 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:07:11.452 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:11.452 [2024-12-05 18:57:28.933697] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:11.452 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:11.452 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:11.452 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:11.452 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:11.452 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:11.452 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@41 -- # break 00:07:11.452 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@45 -- # return 0 00:07:11.452 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:07:11.452 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:11.452 18:57:28 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo '' 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # true 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=0 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 0 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # count=0 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@97 -- # killprocess 71491 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@954 -- # '[' -z 71491 ']' 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@958 -- # kill -0 71491 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@959 -- # uname 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71491 00:07:11.712 killing process with pid 71491 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71491' 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@973 -- # kill 71491 00:07:11.712 [2024-12-05 18:57:29.245549] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:11.712 [2024-12-05 18:57:29.245657] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:11.712 [2024-12-05 18:57:29.245731] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:11.712 18:57:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@978 -- # wait 71491 00:07:11.712 [2024-12-05 18:57:29.245743] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid, state offline 00:07:11.712 [2024-12-05 18:57:29.267754] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:11.972 18:57:29 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@99 -- # return 0 00:07:11.972 00:07:11.972 real 0m2.588s 00:07:11.972 user 0m3.201s 00:07:11.972 sys 0m0.862s 00:07:11.972 18:57:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:11.972 18:57:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:11.972 ************************************ 00:07:11.972 END TEST raid_function_test_raid0 00:07:11.972 ************************************ 00:07:12.232 18:57:29 bdev_raid -- bdev/bdev_raid.sh@960 -- # run_test raid_function_test_concat raid_function_test concat 00:07:12.233 18:57:29 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:07:12.233 18:57:29 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:12.233 18:57:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:12.233 ************************************ 00:07:12.233 START TEST raid_function_test_concat 00:07:12.233 ************************************ 00:07:12.233 18:57:29 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1129 -- # raid_function_test concat 00:07:12.233 18:57:29 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@64 -- # local raid_level=concat 00:07:12.233 18:57:29 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:07:12.233 18:57:29 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:07:12.233 18:57:29 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@69 -- # raid_pid=71606 00:07:12.233 18:57:29 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:12.233 Process raid pid: 71606 00:07:12.233 18:57:29 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 71606' 00:07:12.233 18:57:29 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@71 -- # waitforlisten 71606 00:07:12.233 18:57:29 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@835 -- # '[' -z 71606 ']' 00:07:12.233 18:57:29 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:12.233 18:57:29 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:12.233 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:12.233 18:57:29 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:12.233 18:57:29 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:12.233 18:57:29 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:12.233 [2024-12-05 18:57:29.634848] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:12.233 [2024-12-05 18:57:29.634965] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:12.233 [2024-12-05 18:57:29.769235] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:12.492 [2024-12-05 18:57:29.793206] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:12.492 [2024-12-05 18:57:29.835020] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:12.492 [2024-12-05 18:57:29.835060] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:13.063 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:13.063 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@868 -- # return 0 00:07:13.063 18:57:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:13.064 Base_1 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:13.064 Base_2 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''Base_1 Base_2'\''' -n raid 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:13.064 [2024-12-05 18:57:30.483023] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:13.064 [2024-12-05 18:57:30.484893] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:13.064 [2024-12-05 18:57:30.484974] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:13.064 [2024-12-05 18:57:30.484986] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:13.064 [2024-12-05 18:57:30.485240] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:13.064 [2024-12-05 18:57:30.485424] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:13.064 [2024-12-05 18:57:30.485446] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000001200 00:07:13.064 [2024-12-05 18:57:30.485568] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@12 -- # local i 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:13.064 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:07:13.325 [2024-12-05 18:57:30.694718] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:13.325 /dev/nbd0 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@873 -- # local i 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@877 -- # break 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:07:13.325 1+0 records in 00:07:13.325 1+0 records out 00:07:13.325 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000438526 s, 9.3 MB/s 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@890 -- # size=4096 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@893 -- # return 0 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:13.325 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:13.585 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:13.585 { 00:07:13.585 "nbd_device": "/dev/nbd0", 00:07:13.585 "bdev_name": "raid" 00:07:13.585 } 00:07:13.585 ]' 00:07:13.585 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:13.585 { 00:07:13.585 "nbd_device": "/dev/nbd0", 00:07:13.585 "bdev_name": "raid" 00:07:13.585 } 00:07:13.585 ]' 00:07:13.585 18:57:30 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:13.585 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:07:13.585 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=1 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 1 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # count=1 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@19 -- # local blksize 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # blksize=512 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:07:13.586 4096+0 records in 00:07:13.586 4096+0 records out 00:07:13.586 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0324998 s, 64.5 MB/s 00:07:13.586 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:07:13.845 4096+0 records in 00:07:13.845 4096+0 records out 00:07:13.845 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.18088 s, 11.6 MB/s 00:07:13.845 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:07:13.845 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:13.845 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:07:13.845 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:13.845 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:07:13.845 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:07:13.846 128+0 records in 00:07:13.846 128+0 records out 00:07:13.846 65536 bytes (66 kB, 64 KiB) copied, 0.000982281 s, 66.7 MB/s 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:07:13.846 2035+0 records in 00:07:13.846 2035+0 records out 00:07:13.846 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0124406 s, 83.8 MB/s 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:07:13.846 456+0 records in 00:07:13.846 456+0 records out 00:07:13.846 233472 bytes (233 kB, 228 KiB) copied, 0.0020129 s, 116 MB/s 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@52 -- # return 0 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@51 -- # local i 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:13.846 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:07:14.106 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:14.106 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:14.106 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:14.106 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:14.106 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:14.106 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:14.106 [2024-12-05 18:57:31.565918] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:14.106 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@41 -- # break 00:07:14.106 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@45 -- # return 0 00:07:14.106 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:07:14.106 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:14.106 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo '' 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # true 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=0 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 0 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # count=0 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@97 -- # killprocess 71606 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@954 -- # '[' -z 71606 ']' 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@958 -- # kill -0 71606 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@959 -- # uname 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71606 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:14.366 killing process with pid 71606 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71606' 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@973 -- # kill 71606 00:07:14.366 [2024-12-05 18:57:31.873979] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:14.366 [2024-12-05 18:57:31.874114] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:14.366 18:57:31 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@978 -- # wait 71606 00:07:14.366 [2024-12-05 18:57:31.874173] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:14.366 [2024-12-05 18:57:31.874184] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid, state offline 00:07:14.366 [2024-12-05 18:57:31.896541] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:14.627 18:57:32 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@99 -- # return 0 00:07:14.627 00:07:14.627 real 0m2.559s 00:07:14.627 user 0m3.151s 00:07:14.627 sys 0m0.885s 00:07:14.627 18:57:32 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:14.627 18:57:32 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:14.627 ************************************ 00:07:14.627 END TEST raid_function_test_concat 00:07:14.627 ************************************ 00:07:14.627 18:57:32 bdev_raid -- bdev/bdev_raid.sh@963 -- # run_test raid0_resize_test raid_resize_test 0 00:07:14.627 18:57:32 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:07:14.627 18:57:32 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:14.627 18:57:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:14.627 ************************************ 00:07:14.627 START TEST raid0_resize_test 00:07:14.627 ************************************ 00:07:14.627 18:57:32 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1129 -- # raid_resize_test 0 00:07:14.627 18:57:32 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=0 00:07:14.627 18:57:32 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:07:14.627 18:57:32 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:07:14.627 18:57:32 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:07:14.627 18:57:32 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:07:14.627 18:57:32 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:07:14.627 18:57:32 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:07:14.627 18:57:32 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:07:14.627 18:57:32 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=71717 00:07:14.627 18:57:32 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 71717' 00:07:14.627 18:57:32 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:14.627 Process raid pid: 71717 00:07:14.627 18:57:32 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 71717 00:07:14.886 18:57:32 bdev_raid.raid0_resize_test -- common/autotest_common.sh@835 -- # '[' -z 71717 ']' 00:07:14.886 18:57:32 bdev_raid.raid0_resize_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:14.886 18:57:32 bdev_raid.raid0_resize_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:14.886 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:14.886 18:57:32 bdev_raid.raid0_resize_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:14.887 18:57:32 bdev_raid.raid0_resize_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:14.887 18:57:32 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.887 [2024-12-05 18:57:32.261880] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:14.887 [2024-12-05 18:57:32.262011] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:14.887 [2024-12-05 18:57:32.416309] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:14.887 [2024-12-05 18:57:32.440907] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:15.146 [2024-12-05 18:57:32.483100] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:15.146 [2024-12-05 18:57:32.483155] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:15.716 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:15.716 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@868 -- # return 0 00:07:15.716 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:07:15.716 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.716 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.716 Base_1 00:07:15.716 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.716 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:07:15.716 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.716 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.716 Base_2 00:07:15.716 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.716 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 0 -eq 0 ']' 00:07:15.716 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@350 -- # rpc_cmd bdev_raid_create -z 64 -r 0 -b ''\''Base_1 Base_2'\''' -n Raid 00:07:15.716 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.716 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.717 [2024-12-05 18:57:33.105708] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:15.717 [2024-12-05 18:57:33.107562] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:15.717 [2024-12-05 18:57:33.107616] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:15.717 [2024-12-05 18:57:33.107627] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:15.717 [2024-12-05 18:57:33.107909] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:07:15.717 [2024-12-05 18:57:33.108005] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:15.717 [2024-12-05 18:57:33.108016] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:07:15.717 [2024-12-05 18:57:33.108125] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.717 [2024-12-05 18:57:33.117662] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:15.717 [2024-12-05 18:57:33.117700] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:07:15.717 true 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.717 [2024-12-05 18:57:33.133848] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=131072 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=64 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 0 -eq 0 ']' 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@362 -- # expected_size=64 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 64 '!=' 64 ']' 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.717 [2024-12-05 18:57:33.177536] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:15.717 [2024-12-05 18:57:33.177563] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:07:15.717 [2024-12-05 18:57:33.177588] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 131072 to 262144 00:07:15.717 true 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:07:15.717 [2024-12-05 18:57:33.189724] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=262144 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=128 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 0 -eq 0 ']' 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@378 -- # expected_size=128 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 128 '!=' 128 ']' 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 71717 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@954 -- # '[' -z 71717 ']' 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@958 -- # kill -0 71717 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@959 -- # uname 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:15.717 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71717 00:07:15.977 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:15.977 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:15.977 killing process with pid 71717 00:07:15.977 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71717' 00:07:15.977 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@973 -- # kill 71717 00:07:15.977 [2024-12-05 18:57:33.275566] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:15.977 [2024-12-05 18:57:33.275644] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:15.977 [2024-12-05 18:57:33.275706] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:15.977 [2024-12-05 18:57:33.275716] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:07:15.977 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@978 -- # wait 71717 00:07:15.977 [2024-12-05 18:57:33.277168] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:15.977 18:57:33 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:07:15.977 00:07:15.977 real 0m1.311s 00:07:15.977 user 0m1.477s 00:07:15.977 sys 0m0.285s 00:07:15.977 ************************************ 00:07:15.977 END TEST raid0_resize_test 00:07:15.977 ************************************ 00:07:15.977 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:15.977 18:57:33 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.237 18:57:33 bdev_raid -- bdev/bdev_raid.sh@964 -- # run_test raid1_resize_test raid_resize_test 1 00:07:16.237 18:57:33 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:07:16.237 18:57:33 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:16.237 18:57:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:16.237 ************************************ 00:07:16.237 START TEST raid1_resize_test 00:07:16.237 ************************************ 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1129 -- # raid_resize_test 1 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=1 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=71767 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:16.237 Process raid pid: 71767 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 71767' 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 71767 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- common/autotest_common.sh@835 -- # '[' -z 71767 ']' 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:16.237 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:16.237 18:57:33 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.237 [2024-12-05 18:57:33.640121] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:16.237 [2024-12-05 18:57:33.640270] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:16.237 [2024-12-05 18:57:33.772448] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:16.496 [2024-12-05 18:57:33.796488] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:16.496 [2024-12-05 18:57:33.838286] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:16.496 [2024-12-05 18:57:33.838338] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@868 -- # return 0 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.064 Base_1 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.064 Base_2 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 1 -eq 0 ']' 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@352 -- # rpc_cmd bdev_raid_create -r 1 -b ''\''Base_1 Base_2'\''' -n Raid 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.064 [2024-12-05 18:57:34.480680] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:17.064 [2024-12-05 18:57:34.482507] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:17.064 [2024-12-05 18:57:34.482564] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:17.064 [2024-12-05 18:57:34.482575] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:07:17.064 [2024-12-05 18:57:34.482851] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:07:17.064 [2024-12-05 18:57:34.482998] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:17.064 [2024-12-05 18:57:34.483011] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:07:17.064 [2024-12-05 18:57:34.483129] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.064 [2024-12-05 18:57:34.492642] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:17.064 [2024-12-05 18:57:34.492687] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:07:17.064 true 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.064 [2024-12-05 18:57:34.508837] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=65536 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=32 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 1 -eq 0 ']' 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@364 -- # expected_size=32 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 32 '!=' 32 ']' 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.064 [2024-12-05 18:57:34.552516] bdev_raid.c:2317:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:17.064 [2024-12-05 18:57:34.552542] bdev_raid.c:2330:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:07:17.064 [2024-12-05 18:57:34.552566] bdev_raid.c:2344:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 65536 to 131072 00:07:17.064 true 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.064 [2024-12-05 18:57:34.568683] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=131072 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=64 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 1 -eq 0 ']' 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@380 -- # expected_size=64 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 64 '!=' 64 ']' 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 71767 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@954 -- # '[' -z 71767 ']' 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@958 -- # kill -0 71767 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@959 -- # uname 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:17.064 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71767 00:07:17.324 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:17.324 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:17.324 killing process with pid 71767 00:07:17.324 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71767' 00:07:17.324 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@973 -- # kill 71767 00:07:17.324 [2024-12-05 18:57:34.633474] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:17.324 [2024-12-05 18:57:34.633561] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:17.324 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@978 -- # wait 71767 00:07:17.324 [2024-12-05 18:57:34.633996] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:17.324 [2024-12-05 18:57:34.634025] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:07:17.324 [2024-12-05 18:57:34.635120] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:17.324 18:57:34 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:07:17.324 00:07:17.324 real 0m1.285s 00:07:17.324 user 0m1.440s 00:07:17.324 sys 0m0.282s 00:07:17.324 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:17.324 18:57:34 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.324 ************************************ 00:07:17.324 END TEST raid1_resize_test 00:07:17.324 ************************************ 00:07:17.585 18:57:34 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:07:17.585 18:57:34 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:17.585 18:57:34 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 2 false 00:07:17.585 18:57:34 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:17.585 18:57:34 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:17.585 18:57:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:17.585 ************************************ 00:07:17.585 START TEST raid_state_function_test 00:07:17.585 ************************************ 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 2 false 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=71813 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:17.585 Process raid pid: 71813 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 71813' 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 71813 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 71813 ']' 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:17.585 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:17.585 18:57:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.585 [2024-12-05 18:57:35.006388] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:17.585 [2024-12-05 18:57:35.006536] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:17.844 [2024-12-05 18:57:35.164786] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:17.844 [2024-12-05 18:57:35.189213] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:17.844 [2024-12-05 18:57:35.231098] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:17.844 [2024-12-05 18:57:35.231141] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.408 [2024-12-05 18:57:35.833476] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:18.408 [2024-12-05 18:57:35.833540] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:18.408 [2024-12-05 18:57:35.833553] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:18.408 [2024-12-05 18:57:35.833565] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:18.408 "name": "Existed_Raid", 00:07:18.408 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:18.408 "strip_size_kb": 64, 00:07:18.408 "state": "configuring", 00:07:18.408 "raid_level": "raid0", 00:07:18.408 "superblock": false, 00:07:18.408 "num_base_bdevs": 2, 00:07:18.408 "num_base_bdevs_discovered": 0, 00:07:18.408 "num_base_bdevs_operational": 2, 00:07:18.408 "base_bdevs_list": [ 00:07:18.408 { 00:07:18.408 "name": "BaseBdev1", 00:07:18.408 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:18.408 "is_configured": false, 00:07:18.408 "data_offset": 0, 00:07:18.408 "data_size": 0 00:07:18.408 }, 00:07:18.408 { 00:07:18.408 "name": "BaseBdev2", 00:07:18.408 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:18.408 "is_configured": false, 00:07:18.408 "data_offset": 0, 00:07:18.408 "data_size": 0 00:07:18.408 } 00:07:18.408 ] 00:07:18.408 }' 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:18.408 18:57:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.976 [2024-12-05 18:57:36.256659] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:18.976 [2024-12-05 18:57:36.256709] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.976 [2024-12-05 18:57:36.268648] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:18.976 [2024-12-05 18:57:36.268702] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:18.976 [2024-12-05 18:57:36.268711] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:18.976 [2024-12-05 18:57:36.268730] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.976 [2024-12-05 18:57:36.289592] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:18.976 BaseBdev1 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.976 [ 00:07:18.976 { 00:07:18.976 "name": "BaseBdev1", 00:07:18.976 "aliases": [ 00:07:18.976 "4441db07-c919-449b-97b1-4967f7e63cdf" 00:07:18.976 ], 00:07:18.976 "product_name": "Malloc disk", 00:07:18.976 "block_size": 512, 00:07:18.976 "num_blocks": 65536, 00:07:18.976 "uuid": "4441db07-c919-449b-97b1-4967f7e63cdf", 00:07:18.976 "assigned_rate_limits": { 00:07:18.976 "rw_ios_per_sec": 0, 00:07:18.976 "rw_mbytes_per_sec": 0, 00:07:18.976 "r_mbytes_per_sec": 0, 00:07:18.976 "w_mbytes_per_sec": 0 00:07:18.976 }, 00:07:18.976 "claimed": true, 00:07:18.976 "claim_type": "exclusive_write", 00:07:18.976 "zoned": false, 00:07:18.976 "supported_io_types": { 00:07:18.976 "read": true, 00:07:18.976 "write": true, 00:07:18.976 "unmap": true, 00:07:18.976 "flush": true, 00:07:18.976 "reset": true, 00:07:18.976 "nvme_admin": false, 00:07:18.976 "nvme_io": false, 00:07:18.976 "nvme_io_md": false, 00:07:18.976 "write_zeroes": true, 00:07:18.976 "zcopy": true, 00:07:18.976 "get_zone_info": false, 00:07:18.976 "zone_management": false, 00:07:18.976 "zone_append": false, 00:07:18.976 "compare": false, 00:07:18.976 "compare_and_write": false, 00:07:18.976 "abort": true, 00:07:18.976 "seek_hole": false, 00:07:18.976 "seek_data": false, 00:07:18.976 "copy": true, 00:07:18.976 "nvme_iov_md": false 00:07:18.976 }, 00:07:18.976 "memory_domains": [ 00:07:18.976 { 00:07:18.976 "dma_device_id": "system", 00:07:18.976 "dma_device_type": 1 00:07:18.976 }, 00:07:18.976 { 00:07:18.976 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:18.976 "dma_device_type": 2 00:07:18.976 } 00:07:18.976 ], 00:07:18.976 "driver_specific": {} 00:07:18.976 } 00:07:18.976 ] 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:18.976 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:18.977 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.977 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:18.977 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:18.977 "name": "Existed_Raid", 00:07:18.977 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:18.977 "strip_size_kb": 64, 00:07:18.977 "state": "configuring", 00:07:18.977 "raid_level": "raid0", 00:07:18.977 "superblock": false, 00:07:18.977 "num_base_bdevs": 2, 00:07:18.977 "num_base_bdevs_discovered": 1, 00:07:18.977 "num_base_bdevs_operational": 2, 00:07:18.977 "base_bdevs_list": [ 00:07:18.977 { 00:07:18.977 "name": "BaseBdev1", 00:07:18.977 "uuid": "4441db07-c919-449b-97b1-4967f7e63cdf", 00:07:18.977 "is_configured": true, 00:07:18.977 "data_offset": 0, 00:07:18.977 "data_size": 65536 00:07:18.977 }, 00:07:18.977 { 00:07:18.977 "name": "BaseBdev2", 00:07:18.977 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:18.977 "is_configured": false, 00:07:18.977 "data_offset": 0, 00:07:18.977 "data_size": 0 00:07:18.977 } 00:07:18.977 ] 00:07:18.977 }' 00:07:18.977 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:18.977 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.235 [2024-12-05 18:57:36.720862] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:19.235 [2024-12-05 18:57:36.720905] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.235 [2024-12-05 18:57:36.732869] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:19.235 [2024-12-05 18:57:36.734680] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:19.235 [2024-12-05 18:57:36.734741] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.235 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:19.235 "name": "Existed_Raid", 00:07:19.235 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:19.235 "strip_size_kb": 64, 00:07:19.235 "state": "configuring", 00:07:19.235 "raid_level": "raid0", 00:07:19.235 "superblock": false, 00:07:19.235 "num_base_bdevs": 2, 00:07:19.235 "num_base_bdevs_discovered": 1, 00:07:19.235 "num_base_bdevs_operational": 2, 00:07:19.236 "base_bdevs_list": [ 00:07:19.236 { 00:07:19.236 "name": "BaseBdev1", 00:07:19.236 "uuid": "4441db07-c919-449b-97b1-4967f7e63cdf", 00:07:19.236 "is_configured": true, 00:07:19.236 "data_offset": 0, 00:07:19.236 "data_size": 65536 00:07:19.236 }, 00:07:19.236 { 00:07:19.236 "name": "BaseBdev2", 00:07:19.236 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:19.236 "is_configured": false, 00:07:19.236 "data_offset": 0, 00:07:19.236 "data_size": 0 00:07:19.236 } 00:07:19.236 ] 00:07:19.236 }' 00:07:19.236 18:57:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:19.236 18:57:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.803 [2024-12-05 18:57:37.202934] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:19.803 [2024-12-05 18:57:37.202981] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:19.803 [2024-12-05 18:57:37.202998] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:19.803 [2024-12-05 18:57:37.203270] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:19.803 [2024-12-05 18:57:37.203416] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:19.803 [2024-12-05 18:57:37.203441] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:19.803 [2024-12-05 18:57:37.203680] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:19.803 BaseBdev2 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.803 [ 00:07:19.803 { 00:07:19.803 "name": "BaseBdev2", 00:07:19.803 "aliases": [ 00:07:19.803 "6def0ceb-7aec-4028-bb81-7c209442f2b1" 00:07:19.803 ], 00:07:19.803 "product_name": "Malloc disk", 00:07:19.803 "block_size": 512, 00:07:19.803 "num_blocks": 65536, 00:07:19.803 "uuid": "6def0ceb-7aec-4028-bb81-7c209442f2b1", 00:07:19.803 "assigned_rate_limits": { 00:07:19.803 "rw_ios_per_sec": 0, 00:07:19.803 "rw_mbytes_per_sec": 0, 00:07:19.803 "r_mbytes_per_sec": 0, 00:07:19.803 "w_mbytes_per_sec": 0 00:07:19.803 }, 00:07:19.803 "claimed": true, 00:07:19.803 "claim_type": "exclusive_write", 00:07:19.803 "zoned": false, 00:07:19.803 "supported_io_types": { 00:07:19.803 "read": true, 00:07:19.803 "write": true, 00:07:19.803 "unmap": true, 00:07:19.803 "flush": true, 00:07:19.803 "reset": true, 00:07:19.803 "nvme_admin": false, 00:07:19.803 "nvme_io": false, 00:07:19.803 "nvme_io_md": false, 00:07:19.803 "write_zeroes": true, 00:07:19.803 "zcopy": true, 00:07:19.803 "get_zone_info": false, 00:07:19.803 "zone_management": false, 00:07:19.803 "zone_append": false, 00:07:19.803 "compare": false, 00:07:19.803 "compare_and_write": false, 00:07:19.803 "abort": true, 00:07:19.803 "seek_hole": false, 00:07:19.803 "seek_data": false, 00:07:19.803 "copy": true, 00:07:19.803 "nvme_iov_md": false 00:07:19.803 }, 00:07:19.803 "memory_domains": [ 00:07:19.803 { 00:07:19.803 "dma_device_id": "system", 00:07:19.803 "dma_device_type": 1 00:07:19.803 }, 00:07:19.803 { 00:07:19.803 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:19.803 "dma_device_type": 2 00:07:19.803 } 00:07:19.803 ], 00:07:19.803 "driver_specific": {} 00:07:19.803 } 00:07:19.803 ] 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:19.803 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:19.804 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:19.804 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.804 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:19.804 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:19.804 "name": "Existed_Raid", 00:07:19.804 "uuid": "6410d68d-7a59-482f-9859-94b04f16ad18", 00:07:19.804 "strip_size_kb": 64, 00:07:19.804 "state": "online", 00:07:19.804 "raid_level": "raid0", 00:07:19.804 "superblock": false, 00:07:19.804 "num_base_bdevs": 2, 00:07:19.804 "num_base_bdevs_discovered": 2, 00:07:19.804 "num_base_bdevs_operational": 2, 00:07:19.804 "base_bdevs_list": [ 00:07:19.804 { 00:07:19.804 "name": "BaseBdev1", 00:07:19.804 "uuid": "4441db07-c919-449b-97b1-4967f7e63cdf", 00:07:19.804 "is_configured": true, 00:07:19.804 "data_offset": 0, 00:07:19.804 "data_size": 65536 00:07:19.804 }, 00:07:19.804 { 00:07:19.804 "name": "BaseBdev2", 00:07:19.804 "uuid": "6def0ceb-7aec-4028-bb81-7c209442f2b1", 00:07:19.804 "is_configured": true, 00:07:19.804 "data_offset": 0, 00:07:19.804 "data_size": 65536 00:07:19.804 } 00:07:19.804 ] 00:07:19.804 }' 00:07:19.804 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:19.804 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.374 [2024-12-05 18:57:37.670443] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:20.374 "name": "Existed_Raid", 00:07:20.374 "aliases": [ 00:07:20.374 "6410d68d-7a59-482f-9859-94b04f16ad18" 00:07:20.374 ], 00:07:20.374 "product_name": "Raid Volume", 00:07:20.374 "block_size": 512, 00:07:20.374 "num_blocks": 131072, 00:07:20.374 "uuid": "6410d68d-7a59-482f-9859-94b04f16ad18", 00:07:20.374 "assigned_rate_limits": { 00:07:20.374 "rw_ios_per_sec": 0, 00:07:20.374 "rw_mbytes_per_sec": 0, 00:07:20.374 "r_mbytes_per_sec": 0, 00:07:20.374 "w_mbytes_per_sec": 0 00:07:20.374 }, 00:07:20.374 "claimed": false, 00:07:20.374 "zoned": false, 00:07:20.374 "supported_io_types": { 00:07:20.374 "read": true, 00:07:20.374 "write": true, 00:07:20.374 "unmap": true, 00:07:20.374 "flush": true, 00:07:20.374 "reset": true, 00:07:20.374 "nvme_admin": false, 00:07:20.374 "nvme_io": false, 00:07:20.374 "nvme_io_md": false, 00:07:20.374 "write_zeroes": true, 00:07:20.374 "zcopy": false, 00:07:20.374 "get_zone_info": false, 00:07:20.374 "zone_management": false, 00:07:20.374 "zone_append": false, 00:07:20.374 "compare": false, 00:07:20.374 "compare_and_write": false, 00:07:20.374 "abort": false, 00:07:20.374 "seek_hole": false, 00:07:20.374 "seek_data": false, 00:07:20.374 "copy": false, 00:07:20.374 "nvme_iov_md": false 00:07:20.374 }, 00:07:20.374 "memory_domains": [ 00:07:20.374 { 00:07:20.374 "dma_device_id": "system", 00:07:20.374 "dma_device_type": 1 00:07:20.374 }, 00:07:20.374 { 00:07:20.374 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:20.374 "dma_device_type": 2 00:07:20.374 }, 00:07:20.374 { 00:07:20.374 "dma_device_id": "system", 00:07:20.374 "dma_device_type": 1 00:07:20.374 }, 00:07:20.374 { 00:07:20.374 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:20.374 "dma_device_type": 2 00:07:20.374 } 00:07:20.374 ], 00:07:20.374 "driver_specific": { 00:07:20.374 "raid": { 00:07:20.374 "uuid": "6410d68d-7a59-482f-9859-94b04f16ad18", 00:07:20.374 "strip_size_kb": 64, 00:07:20.374 "state": "online", 00:07:20.374 "raid_level": "raid0", 00:07:20.374 "superblock": false, 00:07:20.374 "num_base_bdevs": 2, 00:07:20.374 "num_base_bdevs_discovered": 2, 00:07:20.374 "num_base_bdevs_operational": 2, 00:07:20.374 "base_bdevs_list": [ 00:07:20.374 { 00:07:20.374 "name": "BaseBdev1", 00:07:20.374 "uuid": "4441db07-c919-449b-97b1-4967f7e63cdf", 00:07:20.374 "is_configured": true, 00:07:20.374 "data_offset": 0, 00:07:20.374 "data_size": 65536 00:07:20.374 }, 00:07:20.374 { 00:07:20.374 "name": "BaseBdev2", 00:07:20.374 "uuid": "6def0ceb-7aec-4028-bb81-7c209442f2b1", 00:07:20.374 "is_configured": true, 00:07:20.374 "data_offset": 0, 00:07:20.374 "data_size": 65536 00:07:20.374 } 00:07:20.374 ] 00:07:20.374 } 00:07:20.374 } 00:07:20.374 }' 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:20.374 BaseBdev2' 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.374 [2024-12-05 18:57:37.873848] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:20.374 [2024-12-05 18:57:37.873878] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:20.374 [2024-12-05 18:57:37.873935] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.374 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.375 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:20.375 "name": "Existed_Raid", 00:07:20.375 "uuid": "6410d68d-7a59-482f-9859-94b04f16ad18", 00:07:20.375 "strip_size_kb": 64, 00:07:20.375 "state": "offline", 00:07:20.375 "raid_level": "raid0", 00:07:20.375 "superblock": false, 00:07:20.375 "num_base_bdevs": 2, 00:07:20.375 "num_base_bdevs_discovered": 1, 00:07:20.375 "num_base_bdevs_operational": 1, 00:07:20.375 "base_bdevs_list": [ 00:07:20.375 { 00:07:20.375 "name": null, 00:07:20.375 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:20.375 "is_configured": false, 00:07:20.375 "data_offset": 0, 00:07:20.375 "data_size": 65536 00:07:20.375 }, 00:07:20.375 { 00:07:20.375 "name": "BaseBdev2", 00:07:20.375 "uuid": "6def0ceb-7aec-4028-bb81-7c209442f2b1", 00:07:20.375 "is_configured": true, 00:07:20.375 "data_offset": 0, 00:07:20.375 "data_size": 65536 00:07:20.375 } 00:07:20.375 ] 00:07:20.375 }' 00:07:20.375 18:57:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:20.375 18:57:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.969 [2024-12-05 18:57:38.312363] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:20.969 [2024-12-05 18:57:38.312421] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 71813 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 71813 ']' 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 71813 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 71813 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:20.969 killing process with pid 71813 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 71813' 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 71813 00:07:20.969 [2024-12-05 18:57:38.403002] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:20.969 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 71813 00:07:20.969 [2024-12-05 18:57:38.403987] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:21.240 00:07:21.240 real 0m3.699s 00:07:21.240 user 0m5.864s 00:07:21.240 sys 0m0.686s 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:21.240 ************************************ 00:07:21.240 END TEST raid_state_function_test 00:07:21.240 ************************************ 00:07:21.240 18:57:38 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 2 true 00:07:21.240 18:57:38 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:21.240 18:57:38 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:21.240 18:57:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:21.240 ************************************ 00:07:21.240 START TEST raid_state_function_test_sb 00:07:21.240 ************************************ 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 2 true 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=72050 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72050' 00:07:21.240 Process raid pid: 72050 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 72050 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 72050 ']' 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:21.240 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:21.240 18:57:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.240 [2024-12-05 18:57:38.770752] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:21.240 [2024-12-05 18:57:38.770886] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:21.501 [2024-12-05 18:57:38.924286] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:21.501 [2024-12-05 18:57:38.948622] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:21.501 [2024-12-05 18:57:38.990332] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:21.501 [2024-12-05 18:57:38.990377] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.072 [2024-12-05 18:57:39.592936] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:22.072 [2024-12-05 18:57:39.593008] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:22.072 [2024-12-05 18:57:39.593024] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:22.072 [2024-12-05 18:57:39.593036] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.072 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.332 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:22.332 "name": "Existed_Raid", 00:07:22.332 "uuid": "34919c11-e30f-49c1-aef4-9b3a931d479d", 00:07:22.332 "strip_size_kb": 64, 00:07:22.332 "state": "configuring", 00:07:22.332 "raid_level": "raid0", 00:07:22.332 "superblock": true, 00:07:22.332 "num_base_bdevs": 2, 00:07:22.332 "num_base_bdevs_discovered": 0, 00:07:22.332 "num_base_bdevs_operational": 2, 00:07:22.332 "base_bdevs_list": [ 00:07:22.332 { 00:07:22.332 "name": "BaseBdev1", 00:07:22.332 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:22.332 "is_configured": false, 00:07:22.332 "data_offset": 0, 00:07:22.332 "data_size": 0 00:07:22.332 }, 00:07:22.332 { 00:07:22.332 "name": "BaseBdev2", 00:07:22.332 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:22.332 "is_configured": false, 00:07:22.332 "data_offset": 0, 00:07:22.332 "data_size": 0 00:07:22.332 } 00:07:22.332 ] 00:07:22.332 }' 00:07:22.332 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:22.332 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.593 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:22.593 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.593 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.593 [2024-12-05 18:57:39.983949] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:22.593 [2024-12-05 18:57:39.984032] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:22.593 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.593 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:22.593 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.593 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.593 [2024-12-05 18:57:39.991958] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:22.593 [2024-12-05 18:57:39.991999] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:22.593 [2024-12-05 18:57:39.992008] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:22.593 [2024-12-05 18:57:39.992028] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:22.593 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.593 18:57:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:22.593 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.593 18:57:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.593 [2024-12-05 18:57:40.008660] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:22.594 BaseBdev1 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.594 [ 00:07:22.594 { 00:07:22.594 "name": "BaseBdev1", 00:07:22.594 "aliases": [ 00:07:22.594 "18abba1a-6277-4fe9-821e-ce7b7ab24709" 00:07:22.594 ], 00:07:22.594 "product_name": "Malloc disk", 00:07:22.594 "block_size": 512, 00:07:22.594 "num_blocks": 65536, 00:07:22.594 "uuid": "18abba1a-6277-4fe9-821e-ce7b7ab24709", 00:07:22.594 "assigned_rate_limits": { 00:07:22.594 "rw_ios_per_sec": 0, 00:07:22.594 "rw_mbytes_per_sec": 0, 00:07:22.594 "r_mbytes_per_sec": 0, 00:07:22.594 "w_mbytes_per_sec": 0 00:07:22.594 }, 00:07:22.594 "claimed": true, 00:07:22.594 "claim_type": "exclusive_write", 00:07:22.594 "zoned": false, 00:07:22.594 "supported_io_types": { 00:07:22.594 "read": true, 00:07:22.594 "write": true, 00:07:22.594 "unmap": true, 00:07:22.594 "flush": true, 00:07:22.594 "reset": true, 00:07:22.594 "nvme_admin": false, 00:07:22.594 "nvme_io": false, 00:07:22.594 "nvme_io_md": false, 00:07:22.594 "write_zeroes": true, 00:07:22.594 "zcopy": true, 00:07:22.594 "get_zone_info": false, 00:07:22.594 "zone_management": false, 00:07:22.594 "zone_append": false, 00:07:22.594 "compare": false, 00:07:22.594 "compare_and_write": false, 00:07:22.594 "abort": true, 00:07:22.594 "seek_hole": false, 00:07:22.594 "seek_data": false, 00:07:22.594 "copy": true, 00:07:22.594 "nvme_iov_md": false 00:07:22.594 }, 00:07:22.594 "memory_domains": [ 00:07:22.594 { 00:07:22.594 "dma_device_id": "system", 00:07:22.594 "dma_device_type": 1 00:07:22.594 }, 00:07:22.594 { 00:07:22.594 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:22.594 "dma_device_type": 2 00:07:22.594 } 00:07:22.594 ], 00:07:22.594 "driver_specific": {} 00:07:22.594 } 00:07:22.594 ] 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:22.594 "name": "Existed_Raid", 00:07:22.594 "uuid": "2408ef08-4467-4d25-9a60-e7780c7bab32", 00:07:22.594 "strip_size_kb": 64, 00:07:22.594 "state": "configuring", 00:07:22.594 "raid_level": "raid0", 00:07:22.594 "superblock": true, 00:07:22.594 "num_base_bdevs": 2, 00:07:22.594 "num_base_bdevs_discovered": 1, 00:07:22.594 "num_base_bdevs_operational": 2, 00:07:22.594 "base_bdevs_list": [ 00:07:22.594 { 00:07:22.594 "name": "BaseBdev1", 00:07:22.594 "uuid": "18abba1a-6277-4fe9-821e-ce7b7ab24709", 00:07:22.594 "is_configured": true, 00:07:22.594 "data_offset": 2048, 00:07:22.594 "data_size": 63488 00:07:22.594 }, 00:07:22.594 { 00:07:22.594 "name": "BaseBdev2", 00:07:22.594 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:22.594 "is_configured": false, 00:07:22.594 "data_offset": 0, 00:07:22.594 "data_size": 0 00:07:22.594 } 00:07:22.594 ] 00:07:22.594 }' 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:22.594 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.164 [2024-12-05 18:57:40.451905] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:23.164 [2024-12-05 18:57:40.451989] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.164 [2024-12-05 18:57:40.463916] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:23.164 [2024-12-05 18:57:40.465760] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:23.164 [2024-12-05 18:57:40.465847] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.164 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.165 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.165 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:23.165 "name": "Existed_Raid", 00:07:23.165 "uuid": "e0baa8f5-22cc-44d7-9dda-be0adbd3a22b", 00:07:23.165 "strip_size_kb": 64, 00:07:23.165 "state": "configuring", 00:07:23.165 "raid_level": "raid0", 00:07:23.165 "superblock": true, 00:07:23.165 "num_base_bdevs": 2, 00:07:23.165 "num_base_bdevs_discovered": 1, 00:07:23.165 "num_base_bdevs_operational": 2, 00:07:23.165 "base_bdevs_list": [ 00:07:23.165 { 00:07:23.165 "name": "BaseBdev1", 00:07:23.165 "uuid": "18abba1a-6277-4fe9-821e-ce7b7ab24709", 00:07:23.165 "is_configured": true, 00:07:23.165 "data_offset": 2048, 00:07:23.165 "data_size": 63488 00:07:23.165 }, 00:07:23.165 { 00:07:23.165 "name": "BaseBdev2", 00:07:23.165 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:23.165 "is_configured": false, 00:07:23.165 "data_offset": 0, 00:07:23.165 "data_size": 0 00:07:23.165 } 00:07:23.165 ] 00:07:23.165 }' 00:07:23.165 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:23.165 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.425 BaseBdev2 00:07:23.425 [2024-12-05 18:57:40.898175] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:23.425 [2024-12-05 18:57:40.898359] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:23.425 [2024-12-05 18:57:40.898375] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:23.425 [2024-12-05 18:57:40.898681] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:23.425 [2024-12-05 18:57:40.898830] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:23.425 [2024-12-05 18:57:40.898848] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:23.425 [2024-12-05 18:57:40.898982] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.425 [ 00:07:23.425 { 00:07:23.425 "name": "BaseBdev2", 00:07:23.425 "aliases": [ 00:07:23.425 "a12519a1-bce1-4400-8775-9fd2d67c42c4" 00:07:23.425 ], 00:07:23.425 "product_name": "Malloc disk", 00:07:23.425 "block_size": 512, 00:07:23.425 "num_blocks": 65536, 00:07:23.425 "uuid": "a12519a1-bce1-4400-8775-9fd2d67c42c4", 00:07:23.425 "assigned_rate_limits": { 00:07:23.425 "rw_ios_per_sec": 0, 00:07:23.425 "rw_mbytes_per_sec": 0, 00:07:23.425 "r_mbytes_per_sec": 0, 00:07:23.425 "w_mbytes_per_sec": 0 00:07:23.425 }, 00:07:23.425 "claimed": true, 00:07:23.425 "claim_type": "exclusive_write", 00:07:23.425 "zoned": false, 00:07:23.425 "supported_io_types": { 00:07:23.425 "read": true, 00:07:23.425 "write": true, 00:07:23.425 "unmap": true, 00:07:23.425 "flush": true, 00:07:23.425 "reset": true, 00:07:23.425 "nvme_admin": false, 00:07:23.425 "nvme_io": false, 00:07:23.425 "nvme_io_md": false, 00:07:23.425 "write_zeroes": true, 00:07:23.425 "zcopy": true, 00:07:23.425 "get_zone_info": false, 00:07:23.425 "zone_management": false, 00:07:23.425 "zone_append": false, 00:07:23.425 "compare": false, 00:07:23.425 "compare_and_write": false, 00:07:23.425 "abort": true, 00:07:23.425 "seek_hole": false, 00:07:23.425 "seek_data": false, 00:07:23.425 "copy": true, 00:07:23.425 "nvme_iov_md": false 00:07:23.425 }, 00:07:23.425 "memory_domains": [ 00:07:23.425 { 00:07:23.425 "dma_device_id": "system", 00:07:23.425 "dma_device_type": 1 00:07:23.425 }, 00:07:23.425 { 00:07:23.425 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:23.425 "dma_device_type": 2 00:07:23.425 } 00:07:23.425 ], 00:07:23.425 "driver_specific": {} 00:07:23.425 } 00:07:23.425 ] 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.425 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.686 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:23.686 "name": "Existed_Raid", 00:07:23.686 "uuid": "e0baa8f5-22cc-44d7-9dda-be0adbd3a22b", 00:07:23.686 "strip_size_kb": 64, 00:07:23.686 "state": "online", 00:07:23.686 "raid_level": "raid0", 00:07:23.686 "superblock": true, 00:07:23.686 "num_base_bdevs": 2, 00:07:23.686 "num_base_bdevs_discovered": 2, 00:07:23.686 "num_base_bdevs_operational": 2, 00:07:23.686 "base_bdevs_list": [ 00:07:23.686 { 00:07:23.686 "name": "BaseBdev1", 00:07:23.686 "uuid": "18abba1a-6277-4fe9-821e-ce7b7ab24709", 00:07:23.686 "is_configured": true, 00:07:23.686 "data_offset": 2048, 00:07:23.686 "data_size": 63488 00:07:23.686 }, 00:07:23.686 { 00:07:23.686 "name": "BaseBdev2", 00:07:23.686 "uuid": "a12519a1-bce1-4400-8775-9fd2d67c42c4", 00:07:23.686 "is_configured": true, 00:07:23.686 "data_offset": 2048, 00:07:23.686 "data_size": 63488 00:07:23.686 } 00:07:23.686 ] 00:07:23.686 }' 00:07:23.686 18:57:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:23.686 18:57:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:23.947 [2024-12-05 18:57:41.309749] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:23.947 "name": "Existed_Raid", 00:07:23.947 "aliases": [ 00:07:23.947 "e0baa8f5-22cc-44d7-9dda-be0adbd3a22b" 00:07:23.947 ], 00:07:23.947 "product_name": "Raid Volume", 00:07:23.947 "block_size": 512, 00:07:23.947 "num_blocks": 126976, 00:07:23.947 "uuid": "e0baa8f5-22cc-44d7-9dda-be0adbd3a22b", 00:07:23.947 "assigned_rate_limits": { 00:07:23.947 "rw_ios_per_sec": 0, 00:07:23.947 "rw_mbytes_per_sec": 0, 00:07:23.947 "r_mbytes_per_sec": 0, 00:07:23.947 "w_mbytes_per_sec": 0 00:07:23.947 }, 00:07:23.947 "claimed": false, 00:07:23.947 "zoned": false, 00:07:23.947 "supported_io_types": { 00:07:23.947 "read": true, 00:07:23.947 "write": true, 00:07:23.947 "unmap": true, 00:07:23.947 "flush": true, 00:07:23.947 "reset": true, 00:07:23.947 "nvme_admin": false, 00:07:23.947 "nvme_io": false, 00:07:23.947 "nvme_io_md": false, 00:07:23.947 "write_zeroes": true, 00:07:23.947 "zcopy": false, 00:07:23.947 "get_zone_info": false, 00:07:23.947 "zone_management": false, 00:07:23.947 "zone_append": false, 00:07:23.947 "compare": false, 00:07:23.947 "compare_and_write": false, 00:07:23.947 "abort": false, 00:07:23.947 "seek_hole": false, 00:07:23.947 "seek_data": false, 00:07:23.947 "copy": false, 00:07:23.947 "nvme_iov_md": false 00:07:23.947 }, 00:07:23.947 "memory_domains": [ 00:07:23.947 { 00:07:23.947 "dma_device_id": "system", 00:07:23.947 "dma_device_type": 1 00:07:23.947 }, 00:07:23.947 { 00:07:23.947 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:23.947 "dma_device_type": 2 00:07:23.947 }, 00:07:23.947 { 00:07:23.947 "dma_device_id": "system", 00:07:23.947 "dma_device_type": 1 00:07:23.947 }, 00:07:23.947 { 00:07:23.947 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:23.947 "dma_device_type": 2 00:07:23.947 } 00:07:23.947 ], 00:07:23.947 "driver_specific": { 00:07:23.947 "raid": { 00:07:23.947 "uuid": "e0baa8f5-22cc-44d7-9dda-be0adbd3a22b", 00:07:23.947 "strip_size_kb": 64, 00:07:23.947 "state": "online", 00:07:23.947 "raid_level": "raid0", 00:07:23.947 "superblock": true, 00:07:23.947 "num_base_bdevs": 2, 00:07:23.947 "num_base_bdevs_discovered": 2, 00:07:23.947 "num_base_bdevs_operational": 2, 00:07:23.947 "base_bdevs_list": [ 00:07:23.947 { 00:07:23.947 "name": "BaseBdev1", 00:07:23.947 "uuid": "18abba1a-6277-4fe9-821e-ce7b7ab24709", 00:07:23.947 "is_configured": true, 00:07:23.947 "data_offset": 2048, 00:07:23.947 "data_size": 63488 00:07:23.947 }, 00:07:23.947 { 00:07:23.947 "name": "BaseBdev2", 00:07:23.947 "uuid": "a12519a1-bce1-4400-8775-9fd2d67c42c4", 00:07:23.947 "is_configured": true, 00:07:23.947 "data_offset": 2048, 00:07:23.947 "data_size": 63488 00:07:23.947 } 00:07:23.947 ] 00:07:23.947 } 00:07:23.947 } 00:07:23.947 }' 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:23.947 BaseBdev2' 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:23.947 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:23.948 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:23.948 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.948 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.948 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:23.948 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:23.948 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:23.948 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:23.948 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:23.948 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:23.948 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:23.948 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.948 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:23.948 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:24.208 [2024-12-05 18:57:41.521175] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:24.208 [2024-12-05 18:57:41.521245] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:24.208 [2024-12-05 18:57:41.521296] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:24.208 "name": "Existed_Raid", 00:07:24.208 "uuid": "e0baa8f5-22cc-44d7-9dda-be0adbd3a22b", 00:07:24.208 "strip_size_kb": 64, 00:07:24.208 "state": "offline", 00:07:24.208 "raid_level": "raid0", 00:07:24.208 "superblock": true, 00:07:24.208 "num_base_bdevs": 2, 00:07:24.208 "num_base_bdevs_discovered": 1, 00:07:24.208 "num_base_bdevs_operational": 1, 00:07:24.208 "base_bdevs_list": [ 00:07:24.208 { 00:07:24.208 "name": null, 00:07:24.208 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:24.208 "is_configured": false, 00:07:24.208 "data_offset": 0, 00:07:24.208 "data_size": 63488 00:07:24.208 }, 00:07:24.208 { 00:07:24.208 "name": "BaseBdev2", 00:07:24.208 "uuid": "a12519a1-bce1-4400-8775-9fd2d67c42c4", 00:07:24.208 "is_configured": true, 00:07:24.208 "data_offset": 2048, 00:07:24.208 "data_size": 63488 00:07:24.208 } 00:07:24.208 ] 00:07:24.208 }' 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:24.208 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:24.468 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:24.468 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:24.468 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:24.468 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:24.468 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.468 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:24.468 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.468 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:24.468 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:24.468 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:24.468 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.468 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:24.468 [2024-12-05 18:57:41.987528] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:24.468 [2024-12-05 18:57:41.987576] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:24.468 18:57:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.468 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:24.468 18:57:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:24.468 18:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:24.468 18:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:24.468 18:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:24.468 18:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:24.469 18:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:24.729 18:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:24.729 18:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:24.729 18:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:24.729 18:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 72050 00:07:24.729 18:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 72050 ']' 00:07:24.729 18:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 72050 00:07:24.729 18:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:07:24.729 18:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:24.729 18:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72050 00:07:24.729 18:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:24.729 18:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:24.729 killing process with pid 72050 00:07:24.729 18:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72050' 00:07:24.729 18:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 72050 00:07:24.729 [2024-12-05 18:57:42.095019] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:24.729 18:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 72050 00:07:24.729 [2024-12-05 18:57:42.095993] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:24.988 18:57:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:24.988 00:07:24.988 real 0m3.623s 00:07:24.988 user 0m5.714s 00:07:24.988 sys 0m0.697s 00:07:24.988 18:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:24.988 18:57:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:24.988 ************************************ 00:07:24.988 END TEST raid_state_function_test_sb 00:07:24.988 ************************************ 00:07:24.988 18:57:42 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 2 00:07:24.988 18:57:42 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:07:24.988 18:57:42 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:24.988 18:57:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:24.988 ************************************ 00:07:24.988 START TEST raid_superblock_test 00:07:24.988 ************************************ 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 2 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=72291 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 72291 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 72291 ']' 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:24.988 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:24.988 18:57:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.988 [2024-12-05 18:57:42.458034] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:24.989 [2024-12-05 18:57:42.458257] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72291 ] 00:07:25.247 [2024-12-05 18:57:42.613009] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:25.247 [2024-12-05 18:57:42.637314] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:25.247 [2024-12-05 18:57:42.679031] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:25.247 [2024-12-05 18:57:42.679146] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.817 malloc1 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.817 [2024-12-05 18:57:43.290137] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:25.817 [2024-12-05 18:57:43.290244] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:25.817 [2024-12-05 18:57:43.290274] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:25.817 [2024-12-05 18:57:43.290289] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:25.817 [2024-12-05 18:57:43.292455] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:25.817 [2024-12-05 18:57:43.292494] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:25.817 pt1 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.817 malloc2 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.817 [2024-12-05 18:57:43.318691] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:25.817 [2024-12-05 18:57:43.318783] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:25.817 [2024-12-05 18:57:43.318838] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:25.817 [2024-12-05 18:57:43.318869] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:25.817 [2024-12-05 18:57:43.321002] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:25.817 [2024-12-05 18:57:43.321073] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:25.817 pt2 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.817 [2024-12-05 18:57:43.330724] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:25.817 [2024-12-05 18:57:43.332525] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:25.817 [2024-12-05 18:57:43.332724] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:25.817 [2024-12-05 18:57:43.332778] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:25.817 [2024-12-05 18:57:43.333051] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:25.817 [2024-12-05 18:57:43.333217] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:25.817 [2024-12-05 18:57:43.333259] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:07:25.817 [2024-12-05 18:57:43.333419] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.817 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.077 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:26.077 "name": "raid_bdev1", 00:07:26.077 "uuid": "7ba1e5d2-6531-41b7-8075-b780f1101bfe", 00:07:26.077 "strip_size_kb": 64, 00:07:26.077 "state": "online", 00:07:26.077 "raid_level": "raid0", 00:07:26.077 "superblock": true, 00:07:26.077 "num_base_bdevs": 2, 00:07:26.077 "num_base_bdevs_discovered": 2, 00:07:26.077 "num_base_bdevs_operational": 2, 00:07:26.077 "base_bdevs_list": [ 00:07:26.077 { 00:07:26.077 "name": "pt1", 00:07:26.077 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:26.077 "is_configured": true, 00:07:26.077 "data_offset": 2048, 00:07:26.077 "data_size": 63488 00:07:26.077 }, 00:07:26.077 { 00:07:26.077 "name": "pt2", 00:07:26.077 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:26.077 "is_configured": true, 00:07:26.077 "data_offset": 2048, 00:07:26.077 "data_size": 63488 00:07:26.077 } 00:07:26.077 ] 00:07:26.077 }' 00:07:26.077 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:26.077 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.336 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:26.336 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:26.336 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:26.336 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:26.336 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:26.336 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:26.336 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:26.336 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:26.336 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.336 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.336 [2024-12-05 18:57:43.754190] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:26.336 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.336 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:26.336 "name": "raid_bdev1", 00:07:26.336 "aliases": [ 00:07:26.336 "7ba1e5d2-6531-41b7-8075-b780f1101bfe" 00:07:26.336 ], 00:07:26.336 "product_name": "Raid Volume", 00:07:26.336 "block_size": 512, 00:07:26.336 "num_blocks": 126976, 00:07:26.336 "uuid": "7ba1e5d2-6531-41b7-8075-b780f1101bfe", 00:07:26.336 "assigned_rate_limits": { 00:07:26.336 "rw_ios_per_sec": 0, 00:07:26.336 "rw_mbytes_per_sec": 0, 00:07:26.336 "r_mbytes_per_sec": 0, 00:07:26.336 "w_mbytes_per_sec": 0 00:07:26.336 }, 00:07:26.336 "claimed": false, 00:07:26.336 "zoned": false, 00:07:26.336 "supported_io_types": { 00:07:26.336 "read": true, 00:07:26.336 "write": true, 00:07:26.336 "unmap": true, 00:07:26.336 "flush": true, 00:07:26.336 "reset": true, 00:07:26.336 "nvme_admin": false, 00:07:26.336 "nvme_io": false, 00:07:26.336 "nvme_io_md": false, 00:07:26.336 "write_zeroes": true, 00:07:26.336 "zcopy": false, 00:07:26.336 "get_zone_info": false, 00:07:26.336 "zone_management": false, 00:07:26.336 "zone_append": false, 00:07:26.336 "compare": false, 00:07:26.336 "compare_and_write": false, 00:07:26.336 "abort": false, 00:07:26.336 "seek_hole": false, 00:07:26.336 "seek_data": false, 00:07:26.336 "copy": false, 00:07:26.336 "nvme_iov_md": false 00:07:26.336 }, 00:07:26.336 "memory_domains": [ 00:07:26.336 { 00:07:26.337 "dma_device_id": "system", 00:07:26.337 "dma_device_type": 1 00:07:26.337 }, 00:07:26.337 { 00:07:26.337 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:26.337 "dma_device_type": 2 00:07:26.337 }, 00:07:26.337 { 00:07:26.337 "dma_device_id": "system", 00:07:26.337 "dma_device_type": 1 00:07:26.337 }, 00:07:26.337 { 00:07:26.337 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:26.337 "dma_device_type": 2 00:07:26.337 } 00:07:26.337 ], 00:07:26.337 "driver_specific": { 00:07:26.337 "raid": { 00:07:26.337 "uuid": "7ba1e5d2-6531-41b7-8075-b780f1101bfe", 00:07:26.337 "strip_size_kb": 64, 00:07:26.337 "state": "online", 00:07:26.337 "raid_level": "raid0", 00:07:26.337 "superblock": true, 00:07:26.337 "num_base_bdevs": 2, 00:07:26.337 "num_base_bdevs_discovered": 2, 00:07:26.337 "num_base_bdevs_operational": 2, 00:07:26.337 "base_bdevs_list": [ 00:07:26.337 { 00:07:26.337 "name": "pt1", 00:07:26.337 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:26.337 "is_configured": true, 00:07:26.337 "data_offset": 2048, 00:07:26.337 "data_size": 63488 00:07:26.337 }, 00:07:26.337 { 00:07:26.337 "name": "pt2", 00:07:26.337 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:26.337 "is_configured": true, 00:07:26.337 "data_offset": 2048, 00:07:26.337 "data_size": 63488 00:07:26.337 } 00:07:26.337 ] 00:07:26.337 } 00:07:26.337 } 00:07:26.337 }' 00:07:26.337 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:26.337 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:26.337 pt2' 00:07:26.337 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.337 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:26.337 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:26.337 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:26.337 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.337 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.337 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.337 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.337 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:26.337 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:26.337 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:26.596 [2024-12-05 18:57:43.949794] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=7ba1e5d2-6531-41b7-8075-b780f1101bfe 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 7ba1e5d2-6531-41b7-8075-b780f1101bfe ']' 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.596 18:57:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.596 [2024-12-05 18:57:43.997477] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:26.596 [2024-12-05 18:57:43.997502] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:26.596 [2024-12-05 18:57:43.997561] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:26.596 [2024-12-05 18:57:43.997610] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:26.596 [2024-12-05 18:57:43.997618] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.596 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.597 [2024-12-05 18:57:44.133280] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:26.597 [2024-12-05 18:57:44.135084] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:26.597 [2024-12-05 18:57:44.135161] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:26.597 [2024-12-05 18:57:44.135203] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:26.597 [2024-12-05 18:57:44.135218] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:26.597 [2024-12-05 18:57:44.135227] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:07:26.597 request: 00:07:26.597 { 00:07:26.597 "name": "raid_bdev1", 00:07:26.597 "raid_level": "raid0", 00:07:26.597 "base_bdevs": [ 00:07:26.597 "malloc1", 00:07:26.597 "malloc2" 00:07:26.597 ], 00:07:26.597 "strip_size_kb": 64, 00:07:26.597 "superblock": false, 00:07:26.597 "method": "bdev_raid_create", 00:07:26.597 "req_id": 1 00:07:26.597 } 00:07:26.597 Got JSON-RPC error response 00:07:26.597 response: 00:07:26.597 { 00:07:26.597 "code": -17, 00:07:26.597 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:26.597 } 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.597 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.856 [2024-12-05 18:57:44.197155] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:26.856 [2024-12-05 18:57:44.197250] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:26.856 [2024-12-05 18:57:44.197290] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:26.856 [2024-12-05 18:57:44.197316] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:26.856 [2024-12-05 18:57:44.199429] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:26.856 [2024-12-05 18:57:44.199496] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:26.856 [2024-12-05 18:57:44.199585] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:26.856 [2024-12-05 18:57:44.199648] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:26.856 pt1 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 2 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:26.856 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:26.856 "name": "raid_bdev1", 00:07:26.856 "uuid": "7ba1e5d2-6531-41b7-8075-b780f1101bfe", 00:07:26.856 "strip_size_kb": 64, 00:07:26.856 "state": "configuring", 00:07:26.856 "raid_level": "raid0", 00:07:26.856 "superblock": true, 00:07:26.856 "num_base_bdevs": 2, 00:07:26.856 "num_base_bdevs_discovered": 1, 00:07:26.856 "num_base_bdevs_operational": 2, 00:07:26.856 "base_bdevs_list": [ 00:07:26.856 { 00:07:26.856 "name": "pt1", 00:07:26.857 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:26.857 "is_configured": true, 00:07:26.857 "data_offset": 2048, 00:07:26.857 "data_size": 63488 00:07:26.857 }, 00:07:26.857 { 00:07:26.857 "name": null, 00:07:26.857 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:26.857 "is_configured": false, 00:07:26.857 "data_offset": 2048, 00:07:26.857 "data_size": 63488 00:07:26.857 } 00:07:26.857 ] 00:07:26.857 }' 00:07:26.857 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:26.857 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.116 [2024-12-05 18:57:44.652357] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:27.116 [2024-12-05 18:57:44.652413] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:27.116 [2024-12-05 18:57:44.652431] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:27.116 [2024-12-05 18:57:44.652439] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:27.116 [2024-12-05 18:57:44.652789] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:27.116 [2024-12-05 18:57:44.652806] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:27.116 [2024-12-05 18:57:44.652863] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:27.116 [2024-12-05 18:57:44.652879] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:27.116 [2024-12-05 18:57:44.652957] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:27.116 [2024-12-05 18:57:44.652970] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:27.116 [2024-12-05 18:57:44.653199] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:27.116 [2024-12-05 18:57:44.653307] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:27.116 [2024-12-05 18:57:44.653319] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:27.116 [2024-12-05 18:57:44.653406] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:27.116 pt2 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.116 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.375 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.375 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:27.375 "name": "raid_bdev1", 00:07:27.375 "uuid": "7ba1e5d2-6531-41b7-8075-b780f1101bfe", 00:07:27.375 "strip_size_kb": 64, 00:07:27.375 "state": "online", 00:07:27.375 "raid_level": "raid0", 00:07:27.375 "superblock": true, 00:07:27.375 "num_base_bdevs": 2, 00:07:27.375 "num_base_bdevs_discovered": 2, 00:07:27.375 "num_base_bdevs_operational": 2, 00:07:27.375 "base_bdevs_list": [ 00:07:27.375 { 00:07:27.375 "name": "pt1", 00:07:27.375 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:27.375 "is_configured": true, 00:07:27.375 "data_offset": 2048, 00:07:27.375 "data_size": 63488 00:07:27.375 }, 00:07:27.375 { 00:07:27.375 "name": "pt2", 00:07:27.375 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:27.375 "is_configured": true, 00:07:27.375 "data_offset": 2048, 00:07:27.375 "data_size": 63488 00:07:27.375 } 00:07:27.375 ] 00:07:27.375 }' 00:07:27.375 18:57:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:27.375 18:57:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:27.635 [2024-12-05 18:57:45.047965] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:27.635 "name": "raid_bdev1", 00:07:27.635 "aliases": [ 00:07:27.635 "7ba1e5d2-6531-41b7-8075-b780f1101bfe" 00:07:27.635 ], 00:07:27.635 "product_name": "Raid Volume", 00:07:27.635 "block_size": 512, 00:07:27.635 "num_blocks": 126976, 00:07:27.635 "uuid": "7ba1e5d2-6531-41b7-8075-b780f1101bfe", 00:07:27.635 "assigned_rate_limits": { 00:07:27.635 "rw_ios_per_sec": 0, 00:07:27.635 "rw_mbytes_per_sec": 0, 00:07:27.635 "r_mbytes_per_sec": 0, 00:07:27.635 "w_mbytes_per_sec": 0 00:07:27.635 }, 00:07:27.635 "claimed": false, 00:07:27.635 "zoned": false, 00:07:27.635 "supported_io_types": { 00:07:27.635 "read": true, 00:07:27.635 "write": true, 00:07:27.635 "unmap": true, 00:07:27.635 "flush": true, 00:07:27.635 "reset": true, 00:07:27.635 "nvme_admin": false, 00:07:27.635 "nvme_io": false, 00:07:27.635 "nvme_io_md": false, 00:07:27.635 "write_zeroes": true, 00:07:27.635 "zcopy": false, 00:07:27.635 "get_zone_info": false, 00:07:27.635 "zone_management": false, 00:07:27.635 "zone_append": false, 00:07:27.635 "compare": false, 00:07:27.635 "compare_and_write": false, 00:07:27.635 "abort": false, 00:07:27.635 "seek_hole": false, 00:07:27.635 "seek_data": false, 00:07:27.635 "copy": false, 00:07:27.635 "nvme_iov_md": false 00:07:27.635 }, 00:07:27.635 "memory_domains": [ 00:07:27.635 { 00:07:27.635 "dma_device_id": "system", 00:07:27.635 "dma_device_type": 1 00:07:27.635 }, 00:07:27.635 { 00:07:27.635 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:27.635 "dma_device_type": 2 00:07:27.635 }, 00:07:27.635 { 00:07:27.635 "dma_device_id": "system", 00:07:27.635 "dma_device_type": 1 00:07:27.635 }, 00:07:27.635 { 00:07:27.635 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:27.635 "dma_device_type": 2 00:07:27.635 } 00:07:27.635 ], 00:07:27.635 "driver_specific": { 00:07:27.635 "raid": { 00:07:27.635 "uuid": "7ba1e5d2-6531-41b7-8075-b780f1101bfe", 00:07:27.635 "strip_size_kb": 64, 00:07:27.635 "state": "online", 00:07:27.635 "raid_level": "raid0", 00:07:27.635 "superblock": true, 00:07:27.635 "num_base_bdevs": 2, 00:07:27.635 "num_base_bdevs_discovered": 2, 00:07:27.635 "num_base_bdevs_operational": 2, 00:07:27.635 "base_bdevs_list": [ 00:07:27.635 { 00:07:27.635 "name": "pt1", 00:07:27.635 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:27.635 "is_configured": true, 00:07:27.635 "data_offset": 2048, 00:07:27.635 "data_size": 63488 00:07:27.635 }, 00:07:27.635 { 00:07:27.635 "name": "pt2", 00:07:27.635 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:27.635 "is_configured": true, 00:07:27.635 "data_offset": 2048, 00:07:27.635 "data_size": 63488 00:07:27.635 } 00:07:27.635 ] 00:07:27.635 } 00:07:27.635 } 00:07:27.635 }' 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:27.635 pt2' 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.635 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.894 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.894 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:27.894 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:27.894 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:27.894 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:27.894 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:27.894 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.894 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.894 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.894 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:27.894 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:27.894 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:27.894 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:27.894 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.895 [2024-12-05 18:57:45.287681] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 7ba1e5d2-6531-41b7-8075-b780f1101bfe '!=' 7ba1e5d2-6531-41b7-8075-b780f1101bfe ']' 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 72291 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 72291 ']' 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 72291 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72291 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:27.895 killing process with pid 72291 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72291' 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 72291 00:07:27.895 [2024-12-05 18:57:45.354929] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:27.895 [2024-12-05 18:57:45.355014] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:27.895 [2024-12-05 18:57:45.355064] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:27.895 [2024-12-05 18:57:45.355074] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:27.895 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 72291 00:07:27.895 [2024-12-05 18:57:45.377681] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:28.154 18:57:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:28.154 00:07:28.154 real 0m3.216s 00:07:28.154 user 0m4.975s 00:07:28.154 sys 0m0.668s 00:07:28.154 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:28.154 ************************************ 00:07:28.154 END TEST raid_superblock_test 00:07:28.154 ************************************ 00:07:28.154 18:57:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:28.154 18:57:45 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 2 read 00:07:28.154 18:57:45 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:28.154 18:57:45 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:28.154 18:57:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:28.154 ************************************ 00:07:28.154 START TEST raid_read_error_test 00:07:28.154 ************************************ 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 2 read 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.hTE7NSCGL7 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72486 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72486 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 72486 ']' 00:07:28.154 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:28.154 18:57:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:28.155 18:57:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:28.155 18:57:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:28.155 18:57:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:28.155 18:57:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:28.414 [2024-12-05 18:57:45.758915] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:28.414 [2024-12-05 18:57:45.759114] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72486 ] 00:07:28.414 [2024-12-05 18:57:45.915634] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:28.414 [2024-12-05 18:57:45.940029] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:28.672 [2024-12-05 18:57:45.981892] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:28.672 [2024-12-05 18:57:45.982019] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.240 BaseBdev1_malloc 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.240 true 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.240 [2024-12-05 18:57:46.605319] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:29.240 [2024-12-05 18:57:46.605371] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:29.240 [2024-12-05 18:57:46.605416] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:29.240 [2024-12-05 18:57:46.605425] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:29.240 [2024-12-05 18:57:46.607583] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:29.240 [2024-12-05 18:57:46.607660] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:29.240 BaseBdev1 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.240 BaseBdev2_malloc 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.240 true 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.240 [2024-12-05 18:57:46.646002] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:29.240 [2024-12-05 18:57:46.646093] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:29.240 [2024-12-05 18:57:46.646130] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:29.240 [2024-12-05 18:57:46.646148] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:29.240 [2024-12-05 18:57:46.648177] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:29.240 [2024-12-05 18:57:46.648215] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:29.240 BaseBdev2 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.240 [2024-12-05 18:57:46.658028] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:29.240 [2024-12-05 18:57:46.659838] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:29.240 [2024-12-05 18:57:46.660031] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:29.240 [2024-12-05 18:57:46.660044] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:29.240 [2024-12-05 18:57:46.660290] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:29.240 [2024-12-05 18:57:46.660439] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:29.240 [2024-12-05 18:57:46.660452] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:29.240 [2024-12-05 18:57:46.660590] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:29.240 "name": "raid_bdev1", 00:07:29.240 "uuid": "b28c72ba-713f-49f3-9b28-b68184fb5dd5", 00:07:29.240 "strip_size_kb": 64, 00:07:29.240 "state": "online", 00:07:29.240 "raid_level": "raid0", 00:07:29.240 "superblock": true, 00:07:29.240 "num_base_bdevs": 2, 00:07:29.240 "num_base_bdevs_discovered": 2, 00:07:29.240 "num_base_bdevs_operational": 2, 00:07:29.240 "base_bdevs_list": [ 00:07:29.240 { 00:07:29.240 "name": "BaseBdev1", 00:07:29.240 "uuid": "b8382ed2-98ce-51c9-ac39-ea6dcce4e324", 00:07:29.240 "is_configured": true, 00:07:29.240 "data_offset": 2048, 00:07:29.240 "data_size": 63488 00:07:29.240 }, 00:07:29.240 { 00:07:29.240 "name": "BaseBdev2", 00:07:29.240 "uuid": "9f180746-f752-5b88-8fc9-39cfee148f6e", 00:07:29.240 "is_configured": true, 00:07:29.240 "data_offset": 2048, 00:07:29.240 "data_size": 63488 00:07:29.240 } 00:07:29.240 ] 00:07:29.240 }' 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:29.240 18:57:46 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.808 18:57:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:29.808 18:57:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:29.808 [2024-12-05 18:57:47.129591] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:30.747 "name": "raid_bdev1", 00:07:30.747 "uuid": "b28c72ba-713f-49f3-9b28-b68184fb5dd5", 00:07:30.747 "strip_size_kb": 64, 00:07:30.747 "state": "online", 00:07:30.747 "raid_level": "raid0", 00:07:30.747 "superblock": true, 00:07:30.747 "num_base_bdevs": 2, 00:07:30.747 "num_base_bdevs_discovered": 2, 00:07:30.747 "num_base_bdevs_operational": 2, 00:07:30.747 "base_bdevs_list": [ 00:07:30.747 { 00:07:30.747 "name": "BaseBdev1", 00:07:30.747 "uuid": "b8382ed2-98ce-51c9-ac39-ea6dcce4e324", 00:07:30.747 "is_configured": true, 00:07:30.747 "data_offset": 2048, 00:07:30.747 "data_size": 63488 00:07:30.747 }, 00:07:30.747 { 00:07:30.747 "name": "BaseBdev2", 00:07:30.747 "uuid": "9f180746-f752-5b88-8fc9-39cfee148f6e", 00:07:30.747 "is_configured": true, 00:07:30.747 "data_offset": 2048, 00:07:30.747 "data_size": 63488 00:07:30.747 } 00:07:30.747 ] 00:07:30.747 }' 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:30.747 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.006 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:31.006 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:31.006 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.006 [2024-12-05 18:57:48.525121] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:31.006 [2024-12-05 18:57:48.525153] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:31.006 [2024-12-05 18:57:48.527735] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:31.006 [2024-12-05 18:57:48.527775] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:31.006 [2024-12-05 18:57:48.527808] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:31.006 [2024-12-05 18:57:48.527818] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:31.006 { 00:07:31.006 "results": [ 00:07:31.006 { 00:07:31.006 "job": "raid_bdev1", 00:07:31.006 "core_mask": "0x1", 00:07:31.007 "workload": "randrw", 00:07:31.007 "percentage": 50, 00:07:31.007 "status": "finished", 00:07:31.007 "queue_depth": 1, 00:07:31.007 "io_size": 131072, 00:07:31.007 "runtime": 1.396444, 00:07:31.007 "iops": 17417.81267276024, 00:07:31.007 "mibps": 2177.22658409503, 00:07:31.007 "io_failed": 1, 00:07:31.007 "io_timeout": 0, 00:07:31.007 "avg_latency_us": 79.05341241134064, 00:07:31.007 "min_latency_us": 25.041048034934498, 00:07:31.007 "max_latency_us": 1337.907423580786 00:07:31.007 } 00:07:31.007 ], 00:07:31.007 "core_count": 1 00:07:31.007 } 00:07:31.007 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:31.007 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72486 00:07:31.007 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 72486 ']' 00:07:31.007 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 72486 00:07:31.007 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:07:31.007 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:31.007 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72486 00:07:31.265 killing process with pid 72486 00:07:31.265 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:31.265 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:31.265 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72486' 00:07:31.265 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 72486 00:07:31.265 [2024-12-05 18:57:48.574604] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:31.265 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 72486 00:07:31.265 [2024-12-05 18:57:48.589685] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:31.265 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.hTE7NSCGL7 00:07:31.265 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:31.265 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:31.265 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:07:31.265 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:31.265 ************************************ 00:07:31.265 END TEST raid_read_error_test 00:07:31.265 ************************************ 00:07:31.265 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:31.265 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:31.265 18:57:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:07:31.265 00:07:31.265 real 0m3.141s 00:07:31.265 user 0m3.984s 00:07:31.265 sys 0m0.471s 00:07:31.265 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:31.265 18:57:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.525 18:57:48 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 2 write 00:07:31.525 18:57:48 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:31.525 18:57:48 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:31.525 18:57:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:31.525 ************************************ 00:07:31.525 START TEST raid_write_error_test 00:07:31.525 ************************************ 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 2 write 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.ztnx1Qgby1 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72615 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72615 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 72615 ']' 00:07:31.525 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:31.525 18:57:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.525 [2024-12-05 18:57:48.974595] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:31.525 [2024-12-05 18:57:48.974753] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72615 ] 00:07:31.784 [2024-12-05 18:57:49.129616] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:31.784 [2024-12-05 18:57:49.153920] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:31.784 [2024-12-05 18:57:49.196488] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:31.784 [2024-12-05 18:57:49.196608] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.353 BaseBdev1_malloc 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.353 true 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.353 [2024-12-05 18:57:49.824509] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:32.353 [2024-12-05 18:57:49.824565] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:32.353 [2024-12-05 18:57:49.824591] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:32.353 [2024-12-05 18:57:49.824600] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:32.353 [2024-12-05 18:57:49.826778] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:32.353 [2024-12-05 18:57:49.826812] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:32.353 BaseBdev1 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.353 BaseBdev2_malloc 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.353 true 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.353 [2024-12-05 18:57:49.864933] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:32.353 [2024-12-05 18:57:49.864980] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:32.353 [2024-12-05 18:57:49.864997] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:32.353 [2024-12-05 18:57:49.865014] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:32.353 [2024-12-05 18:57:49.867071] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:32.353 [2024-12-05 18:57:49.867106] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:32.353 BaseBdev2 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.353 [2024-12-05 18:57:49.876981] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:32.353 [2024-12-05 18:57:49.878817] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:32.353 [2024-12-05 18:57:49.878986] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:32.353 [2024-12-05 18:57:49.879000] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:32.353 [2024-12-05 18:57:49.879239] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:32.353 [2024-12-05 18:57:49.879377] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:32.353 [2024-12-05 18:57:49.879389] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:32.353 [2024-12-05 18:57:49.879516] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.353 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:32.612 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:32.613 "name": "raid_bdev1", 00:07:32.613 "uuid": "103fc840-63e2-4695-a99e-6bdf919c1ca6", 00:07:32.613 "strip_size_kb": 64, 00:07:32.613 "state": "online", 00:07:32.613 "raid_level": "raid0", 00:07:32.613 "superblock": true, 00:07:32.613 "num_base_bdevs": 2, 00:07:32.613 "num_base_bdevs_discovered": 2, 00:07:32.613 "num_base_bdevs_operational": 2, 00:07:32.613 "base_bdevs_list": [ 00:07:32.613 { 00:07:32.613 "name": "BaseBdev1", 00:07:32.613 "uuid": "75d73c44-bcfb-5f08-a289-c06396e1f261", 00:07:32.613 "is_configured": true, 00:07:32.613 "data_offset": 2048, 00:07:32.613 "data_size": 63488 00:07:32.613 }, 00:07:32.613 { 00:07:32.613 "name": "BaseBdev2", 00:07:32.613 "uuid": "5dd48e38-70d2-5719-8197-67fccb184b91", 00:07:32.613 "is_configured": true, 00:07:32.613 "data_offset": 2048, 00:07:32.613 "data_size": 63488 00:07:32.613 } 00:07:32.613 ] 00:07:32.613 }' 00:07:32.613 18:57:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:32.613 18:57:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.874 18:57:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:32.874 18:57:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:32.874 [2024-12-05 18:57:50.388473] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:33.815 "name": "raid_bdev1", 00:07:33.815 "uuid": "103fc840-63e2-4695-a99e-6bdf919c1ca6", 00:07:33.815 "strip_size_kb": 64, 00:07:33.815 "state": "online", 00:07:33.815 "raid_level": "raid0", 00:07:33.815 "superblock": true, 00:07:33.815 "num_base_bdevs": 2, 00:07:33.815 "num_base_bdevs_discovered": 2, 00:07:33.815 "num_base_bdevs_operational": 2, 00:07:33.815 "base_bdevs_list": [ 00:07:33.815 { 00:07:33.815 "name": "BaseBdev1", 00:07:33.815 "uuid": "75d73c44-bcfb-5f08-a289-c06396e1f261", 00:07:33.815 "is_configured": true, 00:07:33.815 "data_offset": 2048, 00:07:33.815 "data_size": 63488 00:07:33.815 }, 00:07:33.815 { 00:07:33.815 "name": "BaseBdev2", 00:07:33.815 "uuid": "5dd48e38-70d2-5719-8197-67fccb184b91", 00:07:33.815 "is_configured": true, 00:07:33.815 "data_offset": 2048, 00:07:33.815 "data_size": 63488 00:07:33.815 } 00:07:33.815 ] 00:07:33.815 }' 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:33.815 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.389 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:34.389 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:34.389 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.389 [2024-12-05 18:57:51.736163] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:34.389 [2024-12-05 18:57:51.736267] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:34.389 [2024-12-05 18:57:51.738911] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:34.389 [2024-12-05 18:57:51.738998] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:34.389 [2024-12-05 18:57:51.739053] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:34.389 [2024-12-05 18:57:51.739093] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:34.389 { 00:07:34.389 "results": [ 00:07:34.389 { 00:07:34.389 "job": "raid_bdev1", 00:07:34.389 "core_mask": "0x1", 00:07:34.389 "workload": "randrw", 00:07:34.389 "percentage": 50, 00:07:34.389 "status": "finished", 00:07:34.389 "queue_depth": 1, 00:07:34.389 "io_size": 131072, 00:07:34.389 "runtime": 1.348591, 00:07:34.389 "iops": 17540.52933765686, 00:07:34.389 "mibps": 2192.5661672071074, 00:07:34.389 "io_failed": 1, 00:07:34.389 "io_timeout": 0, 00:07:34.389 "avg_latency_us": 78.69983652143608, 00:07:34.389 "min_latency_us": 24.929257641921396, 00:07:34.389 "max_latency_us": 1402.2986899563318 00:07:34.389 } 00:07:34.389 ], 00:07:34.389 "core_count": 1 00:07:34.389 } 00:07:34.389 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:34.389 18:57:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72615 00:07:34.389 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 72615 ']' 00:07:34.389 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 72615 00:07:34.389 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:07:34.389 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:34.389 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72615 00:07:34.389 killing process with pid 72615 00:07:34.389 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:34.389 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:34.389 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72615' 00:07:34.389 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 72615 00:07:34.389 [2024-12-05 18:57:51.787997] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:34.389 18:57:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 72615 00:07:34.389 [2024-12-05 18:57:51.803035] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:34.648 18:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.ztnx1Qgby1 00:07:34.648 18:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:34.648 18:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:34.648 18:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:07:34.648 18:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:34.648 18:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:34.648 18:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:34.648 ************************************ 00:07:34.648 END TEST raid_write_error_test 00:07:34.648 ************************************ 00:07:34.648 18:57:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:07:34.648 00:07:34.649 real 0m3.142s 00:07:34.649 user 0m4.001s 00:07:34.649 sys 0m0.482s 00:07:34.649 18:57:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:34.649 18:57:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.649 18:57:52 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:34.649 18:57:52 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 2 false 00:07:34.649 18:57:52 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:34.649 18:57:52 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:34.649 18:57:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:34.649 ************************************ 00:07:34.649 START TEST raid_state_function_test 00:07:34.649 ************************************ 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 2 false 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=72742 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72742' 00:07:34.649 Process raid pid: 72742 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 72742 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 72742 ']' 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:34.649 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:34.649 18:57:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.649 [2024-12-05 18:57:52.178458] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:34.649 [2024-12-05 18:57:52.178592] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:34.909 [2024-12-05 18:57:52.310811] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:34.909 [2024-12-05 18:57:52.334718] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:34.909 [2024-12-05 18:57:52.376378] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:34.909 [2024-12-05 18:57:52.376413] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:35.479 18:57:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:35.479 18:57:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:07:35.479 18:57:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:35.479 18:57:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.479 18:57:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.479 [2024-12-05 18:57:52.998846] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:35.479 [2024-12-05 18:57:52.998905] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:35.479 [2024-12-05 18:57:52.998916] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:35.479 [2024-12-05 18:57:52.998925] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.479 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.738 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:35.739 "name": "Existed_Raid", 00:07:35.739 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:35.739 "strip_size_kb": 64, 00:07:35.739 "state": "configuring", 00:07:35.739 "raid_level": "concat", 00:07:35.739 "superblock": false, 00:07:35.739 "num_base_bdevs": 2, 00:07:35.739 "num_base_bdevs_discovered": 0, 00:07:35.739 "num_base_bdevs_operational": 2, 00:07:35.739 "base_bdevs_list": [ 00:07:35.739 { 00:07:35.739 "name": "BaseBdev1", 00:07:35.739 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:35.739 "is_configured": false, 00:07:35.739 "data_offset": 0, 00:07:35.739 "data_size": 0 00:07:35.739 }, 00:07:35.739 { 00:07:35.739 "name": "BaseBdev2", 00:07:35.739 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:35.739 "is_configured": false, 00:07:35.739 "data_offset": 0, 00:07:35.739 "data_size": 0 00:07:35.739 } 00:07:35.739 ] 00:07:35.739 }' 00:07:35.739 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:35.739 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.998 [2024-12-05 18:57:53.366164] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:35.998 [2024-12-05 18:57:53.366214] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.998 [2024-12-05 18:57:53.378127] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:35.998 [2024-12-05 18:57:53.378169] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:35.998 [2024-12-05 18:57:53.378177] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:35.998 [2024-12-05 18:57:53.378197] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.998 [2024-12-05 18:57:53.399116] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:35.998 BaseBdev1 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.998 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.998 [ 00:07:35.998 { 00:07:35.998 "name": "BaseBdev1", 00:07:35.998 "aliases": [ 00:07:35.998 "2a565c5f-9936-44a9-b475-e6c930d50128" 00:07:35.998 ], 00:07:35.998 "product_name": "Malloc disk", 00:07:35.998 "block_size": 512, 00:07:35.998 "num_blocks": 65536, 00:07:35.998 "uuid": "2a565c5f-9936-44a9-b475-e6c930d50128", 00:07:35.998 "assigned_rate_limits": { 00:07:35.998 "rw_ios_per_sec": 0, 00:07:35.998 "rw_mbytes_per_sec": 0, 00:07:35.998 "r_mbytes_per_sec": 0, 00:07:35.998 "w_mbytes_per_sec": 0 00:07:35.998 }, 00:07:35.998 "claimed": true, 00:07:35.998 "claim_type": "exclusive_write", 00:07:35.998 "zoned": false, 00:07:35.999 "supported_io_types": { 00:07:35.999 "read": true, 00:07:35.999 "write": true, 00:07:35.999 "unmap": true, 00:07:35.999 "flush": true, 00:07:35.999 "reset": true, 00:07:35.999 "nvme_admin": false, 00:07:35.999 "nvme_io": false, 00:07:35.999 "nvme_io_md": false, 00:07:35.999 "write_zeroes": true, 00:07:35.999 "zcopy": true, 00:07:35.999 "get_zone_info": false, 00:07:35.999 "zone_management": false, 00:07:35.999 "zone_append": false, 00:07:35.999 "compare": false, 00:07:35.999 "compare_and_write": false, 00:07:35.999 "abort": true, 00:07:35.999 "seek_hole": false, 00:07:35.999 "seek_data": false, 00:07:35.999 "copy": true, 00:07:35.999 "nvme_iov_md": false 00:07:35.999 }, 00:07:35.999 "memory_domains": [ 00:07:35.999 { 00:07:35.999 "dma_device_id": "system", 00:07:35.999 "dma_device_type": 1 00:07:35.999 }, 00:07:35.999 { 00:07:35.999 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:35.999 "dma_device_type": 2 00:07:35.999 } 00:07:35.999 ], 00:07:35.999 "driver_specific": {} 00:07:35.999 } 00:07:35.999 ] 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:35.999 "name": "Existed_Raid", 00:07:35.999 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:35.999 "strip_size_kb": 64, 00:07:35.999 "state": "configuring", 00:07:35.999 "raid_level": "concat", 00:07:35.999 "superblock": false, 00:07:35.999 "num_base_bdevs": 2, 00:07:35.999 "num_base_bdevs_discovered": 1, 00:07:35.999 "num_base_bdevs_operational": 2, 00:07:35.999 "base_bdevs_list": [ 00:07:35.999 { 00:07:35.999 "name": "BaseBdev1", 00:07:35.999 "uuid": "2a565c5f-9936-44a9-b475-e6c930d50128", 00:07:35.999 "is_configured": true, 00:07:35.999 "data_offset": 0, 00:07:35.999 "data_size": 65536 00:07:35.999 }, 00:07:35.999 { 00:07:35.999 "name": "BaseBdev2", 00:07:35.999 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:35.999 "is_configured": false, 00:07:35.999 "data_offset": 0, 00:07:35.999 "data_size": 0 00:07:35.999 } 00:07:35.999 ] 00:07:35.999 }' 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:35.999 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.569 [2024-12-05 18:57:53.822426] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:36.569 [2024-12-05 18:57:53.822478] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.569 [2024-12-05 18:57:53.834433] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:36.569 [2024-12-05 18:57:53.836312] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:36.569 [2024-12-05 18:57:53.836385] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:36.569 "name": "Existed_Raid", 00:07:36.569 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:36.569 "strip_size_kb": 64, 00:07:36.569 "state": "configuring", 00:07:36.569 "raid_level": "concat", 00:07:36.569 "superblock": false, 00:07:36.569 "num_base_bdevs": 2, 00:07:36.569 "num_base_bdevs_discovered": 1, 00:07:36.569 "num_base_bdevs_operational": 2, 00:07:36.569 "base_bdevs_list": [ 00:07:36.569 { 00:07:36.569 "name": "BaseBdev1", 00:07:36.569 "uuid": "2a565c5f-9936-44a9-b475-e6c930d50128", 00:07:36.569 "is_configured": true, 00:07:36.569 "data_offset": 0, 00:07:36.569 "data_size": 65536 00:07:36.569 }, 00:07:36.569 { 00:07:36.569 "name": "BaseBdev2", 00:07:36.569 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:36.569 "is_configured": false, 00:07:36.569 "data_offset": 0, 00:07:36.569 "data_size": 0 00:07:36.569 } 00:07:36.569 ] 00:07:36.569 }' 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:36.569 18:57:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.830 BaseBdev2 00:07:36.830 [2024-12-05 18:57:54.304488] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:36.830 [2024-12-05 18:57:54.304530] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:36.830 [2024-12-05 18:57:54.304538] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:36.830 [2024-12-05 18:57:54.304821] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:36.830 [2024-12-05 18:57:54.304974] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:36.830 [2024-12-05 18:57:54.304988] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:36.830 [2024-12-05 18:57:54.305195] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.830 [ 00:07:36.830 { 00:07:36.830 "name": "BaseBdev2", 00:07:36.830 "aliases": [ 00:07:36.830 "a30f2dd1-5b37-4c56-a475-763c945d9877" 00:07:36.830 ], 00:07:36.830 "product_name": "Malloc disk", 00:07:36.830 "block_size": 512, 00:07:36.830 "num_blocks": 65536, 00:07:36.830 "uuid": "a30f2dd1-5b37-4c56-a475-763c945d9877", 00:07:36.830 "assigned_rate_limits": { 00:07:36.830 "rw_ios_per_sec": 0, 00:07:36.830 "rw_mbytes_per_sec": 0, 00:07:36.830 "r_mbytes_per_sec": 0, 00:07:36.830 "w_mbytes_per_sec": 0 00:07:36.830 }, 00:07:36.830 "claimed": true, 00:07:36.830 "claim_type": "exclusive_write", 00:07:36.830 "zoned": false, 00:07:36.830 "supported_io_types": { 00:07:36.830 "read": true, 00:07:36.830 "write": true, 00:07:36.830 "unmap": true, 00:07:36.830 "flush": true, 00:07:36.830 "reset": true, 00:07:36.830 "nvme_admin": false, 00:07:36.830 "nvme_io": false, 00:07:36.830 "nvme_io_md": false, 00:07:36.830 "write_zeroes": true, 00:07:36.830 "zcopy": true, 00:07:36.830 "get_zone_info": false, 00:07:36.830 "zone_management": false, 00:07:36.830 "zone_append": false, 00:07:36.830 "compare": false, 00:07:36.830 "compare_and_write": false, 00:07:36.830 "abort": true, 00:07:36.830 "seek_hole": false, 00:07:36.830 "seek_data": false, 00:07:36.830 "copy": true, 00:07:36.830 "nvme_iov_md": false 00:07:36.830 }, 00:07:36.830 "memory_domains": [ 00:07:36.830 { 00:07:36.830 "dma_device_id": "system", 00:07:36.830 "dma_device_type": 1 00:07:36.830 }, 00:07:36.830 { 00:07:36.830 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:36.830 "dma_device_type": 2 00:07:36.830 } 00:07:36.830 ], 00:07:36.830 "driver_specific": {} 00:07:36.830 } 00:07:36.830 ] 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.830 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.090 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:37.090 "name": "Existed_Raid", 00:07:37.090 "uuid": "8a29c282-4673-4641-a2b1-323b0fcf6819", 00:07:37.090 "strip_size_kb": 64, 00:07:37.090 "state": "online", 00:07:37.090 "raid_level": "concat", 00:07:37.090 "superblock": false, 00:07:37.090 "num_base_bdevs": 2, 00:07:37.090 "num_base_bdevs_discovered": 2, 00:07:37.090 "num_base_bdevs_operational": 2, 00:07:37.090 "base_bdevs_list": [ 00:07:37.090 { 00:07:37.090 "name": "BaseBdev1", 00:07:37.090 "uuid": "2a565c5f-9936-44a9-b475-e6c930d50128", 00:07:37.090 "is_configured": true, 00:07:37.090 "data_offset": 0, 00:07:37.090 "data_size": 65536 00:07:37.090 }, 00:07:37.090 { 00:07:37.090 "name": "BaseBdev2", 00:07:37.090 "uuid": "a30f2dd1-5b37-4c56-a475-763c945d9877", 00:07:37.090 "is_configured": true, 00:07:37.090 "data_offset": 0, 00:07:37.090 "data_size": 65536 00:07:37.090 } 00:07:37.090 ] 00:07:37.090 }' 00:07:37.090 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:37.090 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.350 [2024-12-05 18:57:54.748042] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:37.350 "name": "Existed_Raid", 00:07:37.350 "aliases": [ 00:07:37.350 "8a29c282-4673-4641-a2b1-323b0fcf6819" 00:07:37.350 ], 00:07:37.350 "product_name": "Raid Volume", 00:07:37.350 "block_size": 512, 00:07:37.350 "num_blocks": 131072, 00:07:37.350 "uuid": "8a29c282-4673-4641-a2b1-323b0fcf6819", 00:07:37.350 "assigned_rate_limits": { 00:07:37.350 "rw_ios_per_sec": 0, 00:07:37.350 "rw_mbytes_per_sec": 0, 00:07:37.350 "r_mbytes_per_sec": 0, 00:07:37.350 "w_mbytes_per_sec": 0 00:07:37.350 }, 00:07:37.350 "claimed": false, 00:07:37.350 "zoned": false, 00:07:37.350 "supported_io_types": { 00:07:37.350 "read": true, 00:07:37.350 "write": true, 00:07:37.350 "unmap": true, 00:07:37.350 "flush": true, 00:07:37.350 "reset": true, 00:07:37.350 "nvme_admin": false, 00:07:37.350 "nvme_io": false, 00:07:37.350 "nvme_io_md": false, 00:07:37.350 "write_zeroes": true, 00:07:37.350 "zcopy": false, 00:07:37.350 "get_zone_info": false, 00:07:37.350 "zone_management": false, 00:07:37.350 "zone_append": false, 00:07:37.350 "compare": false, 00:07:37.350 "compare_and_write": false, 00:07:37.350 "abort": false, 00:07:37.350 "seek_hole": false, 00:07:37.350 "seek_data": false, 00:07:37.350 "copy": false, 00:07:37.350 "nvme_iov_md": false 00:07:37.350 }, 00:07:37.350 "memory_domains": [ 00:07:37.350 { 00:07:37.350 "dma_device_id": "system", 00:07:37.350 "dma_device_type": 1 00:07:37.350 }, 00:07:37.350 { 00:07:37.350 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:37.350 "dma_device_type": 2 00:07:37.350 }, 00:07:37.350 { 00:07:37.350 "dma_device_id": "system", 00:07:37.350 "dma_device_type": 1 00:07:37.350 }, 00:07:37.350 { 00:07:37.350 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:37.350 "dma_device_type": 2 00:07:37.350 } 00:07:37.350 ], 00:07:37.350 "driver_specific": { 00:07:37.350 "raid": { 00:07:37.350 "uuid": "8a29c282-4673-4641-a2b1-323b0fcf6819", 00:07:37.350 "strip_size_kb": 64, 00:07:37.350 "state": "online", 00:07:37.350 "raid_level": "concat", 00:07:37.350 "superblock": false, 00:07:37.350 "num_base_bdevs": 2, 00:07:37.350 "num_base_bdevs_discovered": 2, 00:07:37.350 "num_base_bdevs_operational": 2, 00:07:37.350 "base_bdevs_list": [ 00:07:37.350 { 00:07:37.350 "name": "BaseBdev1", 00:07:37.350 "uuid": "2a565c5f-9936-44a9-b475-e6c930d50128", 00:07:37.350 "is_configured": true, 00:07:37.350 "data_offset": 0, 00:07:37.350 "data_size": 65536 00:07:37.350 }, 00:07:37.350 { 00:07:37.350 "name": "BaseBdev2", 00:07:37.350 "uuid": "a30f2dd1-5b37-4c56-a475-763c945d9877", 00:07:37.350 "is_configured": true, 00:07:37.350 "data_offset": 0, 00:07:37.350 "data_size": 65536 00:07:37.350 } 00:07:37.350 ] 00:07:37.350 } 00:07:37.350 } 00:07:37.350 }' 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:37.350 BaseBdev2' 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:37.350 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.610 [2024-12-05 18:57:54.963467] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:37.610 [2024-12-05 18:57:54.963548] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:37.610 [2024-12-05 18:57:54.963613] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:37.610 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:37.611 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:37.611 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:37.611 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:37.611 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:37.611 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:37.611 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:37.611 18:57:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:37.611 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:37.611 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.611 18:57:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:37.611 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:37.611 "name": "Existed_Raid", 00:07:37.611 "uuid": "8a29c282-4673-4641-a2b1-323b0fcf6819", 00:07:37.611 "strip_size_kb": 64, 00:07:37.611 "state": "offline", 00:07:37.611 "raid_level": "concat", 00:07:37.611 "superblock": false, 00:07:37.611 "num_base_bdevs": 2, 00:07:37.611 "num_base_bdevs_discovered": 1, 00:07:37.611 "num_base_bdevs_operational": 1, 00:07:37.611 "base_bdevs_list": [ 00:07:37.611 { 00:07:37.611 "name": null, 00:07:37.611 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:37.611 "is_configured": false, 00:07:37.611 "data_offset": 0, 00:07:37.611 "data_size": 65536 00:07:37.611 }, 00:07:37.611 { 00:07:37.611 "name": "BaseBdev2", 00:07:37.611 "uuid": "a30f2dd1-5b37-4c56-a475-763c945d9877", 00:07:37.611 "is_configured": true, 00:07:37.611 "data_offset": 0, 00:07:37.611 "data_size": 65536 00:07:37.611 } 00:07:37.611 ] 00:07:37.611 }' 00:07:37.611 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:37.611 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.179 [2024-12-05 18:57:55.489830] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:38.179 [2024-12-05 18:57:55.489947] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 72742 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 72742 ']' 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 72742 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72742 00:07:38.179 killing process with pid 72742 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72742' 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 72742 00:07:38.179 [2024-12-05 18:57:55.596911] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:38.179 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 72742 00:07:38.179 [2024-12-05 18:57:55.597886] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:38.438 18:57:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:38.439 00:07:38.439 real 0m3.724s 00:07:38.439 user 0m5.867s 00:07:38.439 sys 0m0.733s 00:07:38.439 ************************************ 00:07:38.439 END TEST raid_state_function_test 00:07:38.439 ************************************ 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.439 18:57:55 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 2 true 00:07:38.439 18:57:55 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:38.439 18:57:55 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:38.439 18:57:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:38.439 ************************************ 00:07:38.439 START TEST raid_state_function_test_sb 00:07:38.439 ************************************ 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 2 true 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=72984 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72984' 00:07:38.439 Process raid pid: 72984 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 72984 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 72984 ']' 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:38.439 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:38.439 18:57:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.439 [2024-12-05 18:57:55.969832] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:38.439 [2024-12-05 18:57:55.970045] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:38.698 [2024-12-05 18:57:56.104332] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:38.698 [2024-12-05 18:57:56.128452] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:38.698 [2024-12-05 18:57:56.170543] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:38.698 [2024-12-05 18:57:56.170578] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.267 [2024-12-05 18:57:56.801149] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:39.267 [2024-12-05 18:57:56.801263] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:39.267 [2024-12-05 18:57:56.801277] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:39.267 [2024-12-05 18:57:56.801287] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.267 18:57:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.526 18:57:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.526 18:57:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:39.526 "name": "Existed_Raid", 00:07:39.526 "uuid": "db4f6dfa-8ee2-41db-9b53-b11d7f092b82", 00:07:39.526 "strip_size_kb": 64, 00:07:39.526 "state": "configuring", 00:07:39.527 "raid_level": "concat", 00:07:39.527 "superblock": true, 00:07:39.527 "num_base_bdevs": 2, 00:07:39.527 "num_base_bdevs_discovered": 0, 00:07:39.527 "num_base_bdevs_operational": 2, 00:07:39.527 "base_bdevs_list": [ 00:07:39.527 { 00:07:39.527 "name": "BaseBdev1", 00:07:39.527 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:39.527 "is_configured": false, 00:07:39.527 "data_offset": 0, 00:07:39.527 "data_size": 0 00:07:39.527 }, 00:07:39.527 { 00:07:39.527 "name": "BaseBdev2", 00:07:39.527 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:39.527 "is_configured": false, 00:07:39.527 "data_offset": 0, 00:07:39.527 "data_size": 0 00:07:39.527 } 00:07:39.527 ] 00:07:39.527 }' 00:07:39.527 18:57:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:39.527 18:57:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.788 [2024-12-05 18:57:57.240290] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:39.788 [2024-12-05 18:57:57.240380] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.788 [2024-12-05 18:57:57.252289] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:39.788 [2024-12-05 18:57:57.252367] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:39.788 [2024-12-05 18:57:57.252392] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:39.788 [2024-12-05 18:57:57.252428] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.788 [2024-12-05 18:57:57.273074] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:39.788 BaseBdev1 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.788 [ 00:07:39.788 { 00:07:39.788 "name": "BaseBdev1", 00:07:39.788 "aliases": [ 00:07:39.788 "1b05919b-1003-4f51-9e3d-59c757203116" 00:07:39.788 ], 00:07:39.788 "product_name": "Malloc disk", 00:07:39.788 "block_size": 512, 00:07:39.788 "num_blocks": 65536, 00:07:39.788 "uuid": "1b05919b-1003-4f51-9e3d-59c757203116", 00:07:39.788 "assigned_rate_limits": { 00:07:39.788 "rw_ios_per_sec": 0, 00:07:39.788 "rw_mbytes_per_sec": 0, 00:07:39.788 "r_mbytes_per_sec": 0, 00:07:39.788 "w_mbytes_per_sec": 0 00:07:39.788 }, 00:07:39.788 "claimed": true, 00:07:39.788 "claim_type": "exclusive_write", 00:07:39.788 "zoned": false, 00:07:39.788 "supported_io_types": { 00:07:39.788 "read": true, 00:07:39.788 "write": true, 00:07:39.788 "unmap": true, 00:07:39.788 "flush": true, 00:07:39.788 "reset": true, 00:07:39.788 "nvme_admin": false, 00:07:39.788 "nvme_io": false, 00:07:39.788 "nvme_io_md": false, 00:07:39.788 "write_zeroes": true, 00:07:39.788 "zcopy": true, 00:07:39.788 "get_zone_info": false, 00:07:39.788 "zone_management": false, 00:07:39.788 "zone_append": false, 00:07:39.788 "compare": false, 00:07:39.788 "compare_and_write": false, 00:07:39.788 "abort": true, 00:07:39.788 "seek_hole": false, 00:07:39.788 "seek_data": false, 00:07:39.788 "copy": true, 00:07:39.788 "nvme_iov_md": false 00:07:39.788 }, 00:07:39.788 "memory_domains": [ 00:07:39.788 { 00:07:39.788 "dma_device_id": "system", 00:07:39.788 "dma_device_type": 1 00:07:39.788 }, 00:07:39.788 { 00:07:39.788 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:39.788 "dma_device_type": 2 00:07:39.788 } 00:07:39.788 ], 00:07:39.788 "driver_specific": {} 00:07:39.788 } 00:07:39.788 ] 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:39.788 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.048 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:40.048 "name": "Existed_Raid", 00:07:40.048 "uuid": "72e3dd32-e4ca-48ad-898b-c1b25cc6f336", 00:07:40.048 "strip_size_kb": 64, 00:07:40.048 "state": "configuring", 00:07:40.048 "raid_level": "concat", 00:07:40.048 "superblock": true, 00:07:40.048 "num_base_bdevs": 2, 00:07:40.048 "num_base_bdevs_discovered": 1, 00:07:40.048 "num_base_bdevs_operational": 2, 00:07:40.048 "base_bdevs_list": [ 00:07:40.048 { 00:07:40.048 "name": "BaseBdev1", 00:07:40.048 "uuid": "1b05919b-1003-4f51-9e3d-59c757203116", 00:07:40.048 "is_configured": true, 00:07:40.048 "data_offset": 2048, 00:07:40.048 "data_size": 63488 00:07:40.048 }, 00:07:40.048 { 00:07:40.048 "name": "BaseBdev2", 00:07:40.048 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:40.048 "is_configured": false, 00:07:40.048 "data_offset": 0, 00:07:40.048 "data_size": 0 00:07:40.048 } 00:07:40.048 ] 00:07:40.048 }' 00:07:40.048 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:40.048 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.308 [2024-12-05 18:57:57.768279] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:40.308 [2024-12-05 18:57:57.768375] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.308 [2024-12-05 18:57:57.776293] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:40.308 [2024-12-05 18:57:57.778290] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:40.308 [2024-12-05 18:57:57.778337] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:40.308 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:40.309 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.309 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.309 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.309 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:40.309 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.309 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:40.309 "name": "Existed_Raid", 00:07:40.309 "uuid": "e01ec2d7-2b30-423c-ac2d-f27d53d1866c", 00:07:40.309 "strip_size_kb": 64, 00:07:40.309 "state": "configuring", 00:07:40.309 "raid_level": "concat", 00:07:40.309 "superblock": true, 00:07:40.309 "num_base_bdevs": 2, 00:07:40.309 "num_base_bdevs_discovered": 1, 00:07:40.309 "num_base_bdevs_operational": 2, 00:07:40.309 "base_bdevs_list": [ 00:07:40.309 { 00:07:40.309 "name": "BaseBdev1", 00:07:40.309 "uuid": "1b05919b-1003-4f51-9e3d-59c757203116", 00:07:40.309 "is_configured": true, 00:07:40.309 "data_offset": 2048, 00:07:40.309 "data_size": 63488 00:07:40.309 }, 00:07:40.309 { 00:07:40.309 "name": "BaseBdev2", 00:07:40.309 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:40.309 "is_configured": false, 00:07:40.309 "data_offset": 0, 00:07:40.309 "data_size": 0 00:07:40.309 } 00:07:40.309 ] 00:07:40.309 }' 00:07:40.309 18:57:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:40.309 18:57:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.880 [2024-12-05 18:57:58.202453] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:40.880 [2024-12-05 18:57:58.202739] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:40.880 [2024-12-05 18:57:58.202792] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:40.880 BaseBdev2 00:07:40.880 [2024-12-05 18:57:58.203089] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:40.880 [2024-12-05 18:57:58.203237] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:40.880 [2024-12-05 18:57:58.203296] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:40.880 [2024-12-05 18:57:58.203443] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.880 [ 00:07:40.880 { 00:07:40.880 "name": "BaseBdev2", 00:07:40.880 "aliases": [ 00:07:40.880 "a1222b10-8720-45d0-842a-bb78e9124c67" 00:07:40.880 ], 00:07:40.880 "product_name": "Malloc disk", 00:07:40.880 "block_size": 512, 00:07:40.880 "num_blocks": 65536, 00:07:40.880 "uuid": "a1222b10-8720-45d0-842a-bb78e9124c67", 00:07:40.880 "assigned_rate_limits": { 00:07:40.880 "rw_ios_per_sec": 0, 00:07:40.880 "rw_mbytes_per_sec": 0, 00:07:40.880 "r_mbytes_per_sec": 0, 00:07:40.880 "w_mbytes_per_sec": 0 00:07:40.880 }, 00:07:40.880 "claimed": true, 00:07:40.880 "claim_type": "exclusive_write", 00:07:40.880 "zoned": false, 00:07:40.880 "supported_io_types": { 00:07:40.880 "read": true, 00:07:40.880 "write": true, 00:07:40.880 "unmap": true, 00:07:40.880 "flush": true, 00:07:40.880 "reset": true, 00:07:40.880 "nvme_admin": false, 00:07:40.880 "nvme_io": false, 00:07:40.880 "nvme_io_md": false, 00:07:40.880 "write_zeroes": true, 00:07:40.880 "zcopy": true, 00:07:40.880 "get_zone_info": false, 00:07:40.880 "zone_management": false, 00:07:40.880 "zone_append": false, 00:07:40.880 "compare": false, 00:07:40.880 "compare_and_write": false, 00:07:40.880 "abort": true, 00:07:40.880 "seek_hole": false, 00:07:40.880 "seek_data": false, 00:07:40.880 "copy": true, 00:07:40.880 "nvme_iov_md": false 00:07:40.880 }, 00:07:40.880 "memory_domains": [ 00:07:40.880 { 00:07:40.880 "dma_device_id": "system", 00:07:40.880 "dma_device_type": 1 00:07:40.880 }, 00:07:40.880 { 00:07:40.880 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:40.880 "dma_device_type": 2 00:07:40.880 } 00:07:40.880 ], 00:07:40.880 "driver_specific": {} 00:07:40.880 } 00:07:40.880 ] 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:40.880 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:40.881 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:40.881 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.881 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:40.881 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.881 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:40.881 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:40.881 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:40.881 "name": "Existed_Raid", 00:07:40.881 "uuid": "e01ec2d7-2b30-423c-ac2d-f27d53d1866c", 00:07:40.881 "strip_size_kb": 64, 00:07:40.881 "state": "online", 00:07:40.881 "raid_level": "concat", 00:07:40.881 "superblock": true, 00:07:40.881 "num_base_bdevs": 2, 00:07:40.881 "num_base_bdevs_discovered": 2, 00:07:40.881 "num_base_bdevs_operational": 2, 00:07:40.881 "base_bdevs_list": [ 00:07:40.881 { 00:07:40.881 "name": "BaseBdev1", 00:07:40.881 "uuid": "1b05919b-1003-4f51-9e3d-59c757203116", 00:07:40.881 "is_configured": true, 00:07:40.881 "data_offset": 2048, 00:07:40.881 "data_size": 63488 00:07:40.881 }, 00:07:40.881 { 00:07:40.881 "name": "BaseBdev2", 00:07:40.881 "uuid": "a1222b10-8720-45d0-842a-bb78e9124c67", 00:07:40.881 "is_configured": true, 00:07:40.881 "data_offset": 2048, 00:07:40.881 "data_size": 63488 00:07:40.881 } 00:07:40.881 ] 00:07:40.881 }' 00:07:40.881 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:40.881 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.141 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:41.141 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:41.141 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:41.141 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:41.141 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:41.141 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:41.141 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:41.141 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:41.141 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.141 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.141 [2024-12-05 18:57:58.626000] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:41.141 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.141 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:41.141 "name": "Existed_Raid", 00:07:41.141 "aliases": [ 00:07:41.141 "e01ec2d7-2b30-423c-ac2d-f27d53d1866c" 00:07:41.141 ], 00:07:41.141 "product_name": "Raid Volume", 00:07:41.141 "block_size": 512, 00:07:41.141 "num_blocks": 126976, 00:07:41.141 "uuid": "e01ec2d7-2b30-423c-ac2d-f27d53d1866c", 00:07:41.141 "assigned_rate_limits": { 00:07:41.141 "rw_ios_per_sec": 0, 00:07:41.141 "rw_mbytes_per_sec": 0, 00:07:41.141 "r_mbytes_per_sec": 0, 00:07:41.141 "w_mbytes_per_sec": 0 00:07:41.141 }, 00:07:41.141 "claimed": false, 00:07:41.141 "zoned": false, 00:07:41.141 "supported_io_types": { 00:07:41.141 "read": true, 00:07:41.141 "write": true, 00:07:41.141 "unmap": true, 00:07:41.141 "flush": true, 00:07:41.141 "reset": true, 00:07:41.141 "nvme_admin": false, 00:07:41.141 "nvme_io": false, 00:07:41.141 "nvme_io_md": false, 00:07:41.141 "write_zeroes": true, 00:07:41.141 "zcopy": false, 00:07:41.141 "get_zone_info": false, 00:07:41.141 "zone_management": false, 00:07:41.141 "zone_append": false, 00:07:41.141 "compare": false, 00:07:41.141 "compare_and_write": false, 00:07:41.141 "abort": false, 00:07:41.141 "seek_hole": false, 00:07:41.141 "seek_data": false, 00:07:41.141 "copy": false, 00:07:41.141 "nvme_iov_md": false 00:07:41.141 }, 00:07:41.141 "memory_domains": [ 00:07:41.141 { 00:07:41.141 "dma_device_id": "system", 00:07:41.141 "dma_device_type": 1 00:07:41.141 }, 00:07:41.141 { 00:07:41.141 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:41.141 "dma_device_type": 2 00:07:41.141 }, 00:07:41.141 { 00:07:41.141 "dma_device_id": "system", 00:07:41.141 "dma_device_type": 1 00:07:41.141 }, 00:07:41.141 { 00:07:41.141 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:41.141 "dma_device_type": 2 00:07:41.141 } 00:07:41.141 ], 00:07:41.141 "driver_specific": { 00:07:41.141 "raid": { 00:07:41.141 "uuid": "e01ec2d7-2b30-423c-ac2d-f27d53d1866c", 00:07:41.141 "strip_size_kb": 64, 00:07:41.141 "state": "online", 00:07:41.141 "raid_level": "concat", 00:07:41.141 "superblock": true, 00:07:41.141 "num_base_bdevs": 2, 00:07:41.141 "num_base_bdevs_discovered": 2, 00:07:41.141 "num_base_bdevs_operational": 2, 00:07:41.141 "base_bdevs_list": [ 00:07:41.141 { 00:07:41.141 "name": "BaseBdev1", 00:07:41.141 "uuid": "1b05919b-1003-4f51-9e3d-59c757203116", 00:07:41.141 "is_configured": true, 00:07:41.141 "data_offset": 2048, 00:07:41.141 "data_size": 63488 00:07:41.141 }, 00:07:41.141 { 00:07:41.141 "name": "BaseBdev2", 00:07:41.141 "uuid": "a1222b10-8720-45d0-842a-bb78e9124c67", 00:07:41.141 "is_configured": true, 00:07:41.141 "data_offset": 2048, 00:07:41.141 "data_size": 63488 00:07:41.141 } 00:07:41.141 ] 00:07:41.141 } 00:07:41.141 } 00:07:41.141 }' 00:07:41.141 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:41.402 BaseBdev2' 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.402 [2024-12-05 18:57:58.849451] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:41.402 [2024-12-05 18:57:58.849523] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:41.402 [2024-12-05 18:57:58.849627] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:41.402 "name": "Existed_Raid", 00:07:41.402 "uuid": "e01ec2d7-2b30-423c-ac2d-f27d53d1866c", 00:07:41.402 "strip_size_kb": 64, 00:07:41.402 "state": "offline", 00:07:41.402 "raid_level": "concat", 00:07:41.402 "superblock": true, 00:07:41.402 "num_base_bdevs": 2, 00:07:41.402 "num_base_bdevs_discovered": 1, 00:07:41.402 "num_base_bdevs_operational": 1, 00:07:41.402 "base_bdevs_list": [ 00:07:41.402 { 00:07:41.402 "name": null, 00:07:41.402 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:41.402 "is_configured": false, 00:07:41.402 "data_offset": 0, 00:07:41.402 "data_size": 63488 00:07:41.402 }, 00:07:41.402 { 00:07:41.402 "name": "BaseBdev2", 00:07:41.402 "uuid": "a1222b10-8720-45d0-842a-bb78e9124c67", 00:07:41.402 "is_configured": true, 00:07:41.402 "data_offset": 2048, 00:07:41.402 "data_size": 63488 00:07:41.402 } 00:07:41.402 ] 00:07:41.402 }' 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:41.402 18:57:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.973 [2024-12-05 18:57:59.359914] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:41.973 [2024-12-05 18:57:59.360009] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:41.973 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 72984 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 72984 ']' 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 72984 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 72984 00:07:41.974 killing process with pid 72984 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 72984' 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 72984 00:07:41.974 [2024-12-05 18:57:59.464683] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:41.974 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 72984 00:07:41.974 [2024-12-05 18:57:59.465619] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:42.234 ************************************ 00:07:42.234 END TEST raid_state_function_test_sb 00:07:42.234 ************************************ 00:07:42.234 18:57:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:42.234 00:07:42.234 real 0m3.794s 00:07:42.234 user 0m6.040s 00:07:42.234 sys 0m0.708s 00:07:42.234 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:42.234 18:57:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:42.234 18:57:59 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 2 00:07:42.234 18:57:59 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:07:42.234 18:57:59 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:42.234 18:57:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:42.234 ************************************ 00:07:42.234 START TEST raid_superblock_test 00:07:42.234 ************************************ 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 2 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=73216 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 73216 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 73216 ']' 00:07:42.234 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:42.234 18:57:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.494 [2024-12-05 18:57:59.832846] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:42.494 [2024-12-05 18:57:59.833044] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73216 ] 00:07:42.494 [2024-12-05 18:57:59.988048] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:42.494 [2024-12-05 18:58:00.012365] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:42.753 [2024-12-05 18:58:00.054205] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:42.753 [2024-12-05 18:58:00.054328] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.323 malloc1 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.323 [2024-12-05 18:58:00.665573] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:43.323 [2024-12-05 18:58:00.665633] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:43.323 [2024-12-05 18:58:00.665670] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:43.323 [2024-12-05 18:58:00.665692] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:43.323 [2024-12-05 18:58:00.667796] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:43.323 [2024-12-05 18:58:00.667836] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:43.323 pt1 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.323 malloc2 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.323 [2024-12-05 18:58:00.693933] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:43.323 [2024-12-05 18:58:00.694027] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:43.323 [2024-12-05 18:58:00.694080] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:43.323 [2024-12-05 18:58:00.694109] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:43.323 [2024-12-05 18:58:00.696156] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:43.323 [2024-12-05 18:58:00.696227] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:43.323 pt2 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.323 [2024-12-05 18:58:00.705956] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:43.323 [2024-12-05 18:58:00.707846] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:43.323 [2024-12-05 18:58:00.708021] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:43.323 [2024-12-05 18:58:00.708070] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:43.323 [2024-12-05 18:58:00.708337] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:43.323 [2024-12-05 18:58:00.708505] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:43.323 [2024-12-05 18:58:00.708548] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:07:43.323 [2024-12-05 18:58:00.708711] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:43.323 "name": "raid_bdev1", 00:07:43.323 "uuid": "b0d4fc0d-180c-4d39-a065-d0d6db95225f", 00:07:43.323 "strip_size_kb": 64, 00:07:43.323 "state": "online", 00:07:43.323 "raid_level": "concat", 00:07:43.323 "superblock": true, 00:07:43.323 "num_base_bdevs": 2, 00:07:43.323 "num_base_bdevs_discovered": 2, 00:07:43.323 "num_base_bdevs_operational": 2, 00:07:43.323 "base_bdevs_list": [ 00:07:43.323 { 00:07:43.323 "name": "pt1", 00:07:43.323 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:43.323 "is_configured": true, 00:07:43.323 "data_offset": 2048, 00:07:43.323 "data_size": 63488 00:07:43.323 }, 00:07:43.323 { 00:07:43.323 "name": "pt2", 00:07:43.323 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:43.323 "is_configured": true, 00:07:43.323 "data_offset": 2048, 00:07:43.323 "data_size": 63488 00:07:43.323 } 00:07:43.323 ] 00:07:43.323 }' 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:43.323 18:58:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.893 [2024-12-05 18:58:01.157437] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:43.893 "name": "raid_bdev1", 00:07:43.893 "aliases": [ 00:07:43.893 "b0d4fc0d-180c-4d39-a065-d0d6db95225f" 00:07:43.893 ], 00:07:43.893 "product_name": "Raid Volume", 00:07:43.893 "block_size": 512, 00:07:43.893 "num_blocks": 126976, 00:07:43.893 "uuid": "b0d4fc0d-180c-4d39-a065-d0d6db95225f", 00:07:43.893 "assigned_rate_limits": { 00:07:43.893 "rw_ios_per_sec": 0, 00:07:43.893 "rw_mbytes_per_sec": 0, 00:07:43.893 "r_mbytes_per_sec": 0, 00:07:43.893 "w_mbytes_per_sec": 0 00:07:43.893 }, 00:07:43.893 "claimed": false, 00:07:43.893 "zoned": false, 00:07:43.893 "supported_io_types": { 00:07:43.893 "read": true, 00:07:43.893 "write": true, 00:07:43.893 "unmap": true, 00:07:43.893 "flush": true, 00:07:43.893 "reset": true, 00:07:43.893 "nvme_admin": false, 00:07:43.893 "nvme_io": false, 00:07:43.893 "nvme_io_md": false, 00:07:43.893 "write_zeroes": true, 00:07:43.893 "zcopy": false, 00:07:43.893 "get_zone_info": false, 00:07:43.893 "zone_management": false, 00:07:43.893 "zone_append": false, 00:07:43.893 "compare": false, 00:07:43.893 "compare_and_write": false, 00:07:43.893 "abort": false, 00:07:43.893 "seek_hole": false, 00:07:43.893 "seek_data": false, 00:07:43.893 "copy": false, 00:07:43.893 "nvme_iov_md": false 00:07:43.893 }, 00:07:43.893 "memory_domains": [ 00:07:43.893 { 00:07:43.893 "dma_device_id": "system", 00:07:43.893 "dma_device_type": 1 00:07:43.893 }, 00:07:43.893 { 00:07:43.893 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:43.893 "dma_device_type": 2 00:07:43.893 }, 00:07:43.893 { 00:07:43.893 "dma_device_id": "system", 00:07:43.893 "dma_device_type": 1 00:07:43.893 }, 00:07:43.893 { 00:07:43.893 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:43.893 "dma_device_type": 2 00:07:43.893 } 00:07:43.893 ], 00:07:43.893 "driver_specific": { 00:07:43.893 "raid": { 00:07:43.893 "uuid": "b0d4fc0d-180c-4d39-a065-d0d6db95225f", 00:07:43.893 "strip_size_kb": 64, 00:07:43.893 "state": "online", 00:07:43.893 "raid_level": "concat", 00:07:43.893 "superblock": true, 00:07:43.893 "num_base_bdevs": 2, 00:07:43.893 "num_base_bdevs_discovered": 2, 00:07:43.893 "num_base_bdevs_operational": 2, 00:07:43.893 "base_bdevs_list": [ 00:07:43.893 { 00:07:43.893 "name": "pt1", 00:07:43.893 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:43.893 "is_configured": true, 00:07:43.893 "data_offset": 2048, 00:07:43.893 "data_size": 63488 00:07:43.893 }, 00:07:43.893 { 00:07:43.893 "name": "pt2", 00:07:43.893 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:43.893 "is_configured": true, 00:07:43.893 "data_offset": 2048, 00:07:43.893 "data_size": 63488 00:07:43.893 } 00:07:43.893 ] 00:07:43.893 } 00:07:43.893 } 00:07:43.893 }' 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:43.893 pt2' 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.893 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.894 [2024-12-05 18:58:01.357014] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=b0d4fc0d-180c-4d39-a065-d0d6db95225f 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z b0d4fc0d-180c-4d39-a065-d0d6db95225f ']' 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.894 [2024-12-05 18:58:01.404724] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:43.894 [2024-12-05 18:58:01.404751] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:43.894 [2024-12-05 18:58:01.404833] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:43.894 [2024-12-05 18:58:01.404882] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:43.894 [2024-12-05 18:58:01.404895] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.894 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.154 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.154 [2024-12-05 18:58:01.528512] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:44.154 [2024-12-05 18:58:01.530359] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:44.154 [2024-12-05 18:58:01.530440] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:44.154 [2024-12-05 18:58:01.530484] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:44.154 [2024-12-05 18:58:01.530499] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:44.154 [2024-12-05 18:58:01.530508] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:07:44.154 request: 00:07:44.154 { 00:07:44.154 "name": "raid_bdev1", 00:07:44.154 "raid_level": "concat", 00:07:44.154 "base_bdevs": [ 00:07:44.154 "malloc1", 00:07:44.154 "malloc2" 00:07:44.154 ], 00:07:44.154 "strip_size_kb": 64, 00:07:44.154 "superblock": false, 00:07:44.154 "method": "bdev_raid_create", 00:07:44.154 "req_id": 1 00:07:44.154 } 00:07:44.154 Got JSON-RPC error response 00:07:44.154 response: 00:07:44.154 { 00:07:44.154 "code": -17, 00:07:44.154 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:44.154 } 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.155 [2024-12-05 18:58:01.596386] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:44.155 [2024-12-05 18:58:01.596480] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:44.155 [2024-12-05 18:58:01.596520] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:44.155 [2024-12-05 18:58:01.596558] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:44.155 [2024-12-05 18:58:01.598714] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:44.155 [2024-12-05 18:58:01.598776] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:44.155 [2024-12-05 18:58:01.598856] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:44.155 [2024-12-05 18:58:01.598911] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:44.155 pt1 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 2 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:44.155 "name": "raid_bdev1", 00:07:44.155 "uuid": "b0d4fc0d-180c-4d39-a065-d0d6db95225f", 00:07:44.155 "strip_size_kb": 64, 00:07:44.155 "state": "configuring", 00:07:44.155 "raid_level": "concat", 00:07:44.155 "superblock": true, 00:07:44.155 "num_base_bdevs": 2, 00:07:44.155 "num_base_bdevs_discovered": 1, 00:07:44.155 "num_base_bdevs_operational": 2, 00:07:44.155 "base_bdevs_list": [ 00:07:44.155 { 00:07:44.155 "name": "pt1", 00:07:44.155 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:44.155 "is_configured": true, 00:07:44.155 "data_offset": 2048, 00:07:44.155 "data_size": 63488 00:07:44.155 }, 00:07:44.155 { 00:07:44.155 "name": null, 00:07:44.155 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:44.155 "is_configured": false, 00:07:44.155 "data_offset": 2048, 00:07:44.155 "data_size": 63488 00:07:44.155 } 00:07:44.155 ] 00:07:44.155 }' 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:44.155 18:58:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.726 [2024-12-05 18:58:02.027644] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:44.726 [2024-12-05 18:58:02.027752] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:44.726 [2024-12-05 18:58:02.027791] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:44.726 [2024-12-05 18:58:02.027817] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:44.726 [2024-12-05 18:58:02.028223] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:44.726 [2024-12-05 18:58:02.028283] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:44.726 [2024-12-05 18:58:02.028387] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:44.726 [2024-12-05 18:58:02.028436] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:44.726 [2024-12-05 18:58:02.028551] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:44.726 [2024-12-05 18:58:02.028588] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:44.726 [2024-12-05 18:58:02.028864] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:44.726 [2024-12-05 18:58:02.029009] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:44.726 [2024-12-05 18:58:02.029053] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:44.726 [2024-12-05 18:58:02.029189] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:44.726 pt2 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.726 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.727 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.727 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:44.727 "name": "raid_bdev1", 00:07:44.727 "uuid": "b0d4fc0d-180c-4d39-a065-d0d6db95225f", 00:07:44.727 "strip_size_kb": 64, 00:07:44.727 "state": "online", 00:07:44.727 "raid_level": "concat", 00:07:44.727 "superblock": true, 00:07:44.727 "num_base_bdevs": 2, 00:07:44.727 "num_base_bdevs_discovered": 2, 00:07:44.727 "num_base_bdevs_operational": 2, 00:07:44.727 "base_bdevs_list": [ 00:07:44.727 { 00:07:44.727 "name": "pt1", 00:07:44.727 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:44.727 "is_configured": true, 00:07:44.727 "data_offset": 2048, 00:07:44.727 "data_size": 63488 00:07:44.727 }, 00:07:44.727 { 00:07:44.727 "name": "pt2", 00:07:44.727 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:44.727 "is_configured": true, 00:07:44.727 "data_offset": 2048, 00:07:44.727 "data_size": 63488 00:07:44.727 } 00:07:44.727 ] 00:07:44.727 }' 00:07:44.727 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:44.727 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.987 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:44.987 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:44.987 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:44.987 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:44.987 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:44.987 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:44.987 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:44.987 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:44.987 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:44.987 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.987 [2024-12-05 18:58:02.487085] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:44.987 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:44.987 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:44.987 "name": "raid_bdev1", 00:07:44.987 "aliases": [ 00:07:44.987 "b0d4fc0d-180c-4d39-a065-d0d6db95225f" 00:07:44.987 ], 00:07:44.987 "product_name": "Raid Volume", 00:07:44.987 "block_size": 512, 00:07:44.987 "num_blocks": 126976, 00:07:44.987 "uuid": "b0d4fc0d-180c-4d39-a065-d0d6db95225f", 00:07:44.987 "assigned_rate_limits": { 00:07:44.987 "rw_ios_per_sec": 0, 00:07:44.987 "rw_mbytes_per_sec": 0, 00:07:44.987 "r_mbytes_per_sec": 0, 00:07:44.987 "w_mbytes_per_sec": 0 00:07:44.987 }, 00:07:44.987 "claimed": false, 00:07:44.987 "zoned": false, 00:07:44.987 "supported_io_types": { 00:07:44.987 "read": true, 00:07:44.987 "write": true, 00:07:44.987 "unmap": true, 00:07:44.987 "flush": true, 00:07:44.987 "reset": true, 00:07:44.987 "nvme_admin": false, 00:07:44.987 "nvme_io": false, 00:07:44.987 "nvme_io_md": false, 00:07:44.987 "write_zeroes": true, 00:07:44.987 "zcopy": false, 00:07:44.987 "get_zone_info": false, 00:07:44.987 "zone_management": false, 00:07:44.987 "zone_append": false, 00:07:44.987 "compare": false, 00:07:44.987 "compare_and_write": false, 00:07:44.987 "abort": false, 00:07:44.987 "seek_hole": false, 00:07:44.987 "seek_data": false, 00:07:44.987 "copy": false, 00:07:44.987 "nvme_iov_md": false 00:07:44.987 }, 00:07:44.987 "memory_domains": [ 00:07:44.987 { 00:07:44.987 "dma_device_id": "system", 00:07:44.987 "dma_device_type": 1 00:07:44.987 }, 00:07:44.987 { 00:07:44.987 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:44.987 "dma_device_type": 2 00:07:44.987 }, 00:07:44.987 { 00:07:44.988 "dma_device_id": "system", 00:07:44.988 "dma_device_type": 1 00:07:44.988 }, 00:07:44.988 { 00:07:44.988 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:44.988 "dma_device_type": 2 00:07:44.988 } 00:07:44.988 ], 00:07:44.988 "driver_specific": { 00:07:44.988 "raid": { 00:07:44.988 "uuid": "b0d4fc0d-180c-4d39-a065-d0d6db95225f", 00:07:44.988 "strip_size_kb": 64, 00:07:44.988 "state": "online", 00:07:44.988 "raid_level": "concat", 00:07:44.988 "superblock": true, 00:07:44.988 "num_base_bdevs": 2, 00:07:44.988 "num_base_bdevs_discovered": 2, 00:07:44.988 "num_base_bdevs_operational": 2, 00:07:44.988 "base_bdevs_list": [ 00:07:44.988 { 00:07:44.988 "name": "pt1", 00:07:44.988 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:44.988 "is_configured": true, 00:07:44.988 "data_offset": 2048, 00:07:44.988 "data_size": 63488 00:07:44.988 }, 00:07:44.988 { 00:07:44.988 "name": "pt2", 00:07:44.988 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:44.988 "is_configured": true, 00:07:44.988 "data_offset": 2048, 00:07:44.988 "data_size": 63488 00:07:44.988 } 00:07:44.988 ] 00:07:44.988 } 00:07:44.988 } 00:07:44.988 }' 00:07:44.988 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:45.248 pt2' 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.248 [2024-12-05 18:58:02.706714] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' b0d4fc0d-180c-4d39-a065-d0d6db95225f '!=' b0d4fc0d-180c-4d39-a065-d0d6db95225f ']' 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 73216 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 73216 ']' 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 73216 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73216 00:07:45.248 killing process with pid 73216 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73216' 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 73216 00:07:45.248 [2024-12-05 18:58:02.784198] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:45.248 [2024-12-05 18:58:02.784282] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:45.248 [2024-12-05 18:58:02.784332] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:45.248 [2024-12-05 18:58:02.784341] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:45.248 18:58:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 73216 00:07:45.508 [2024-12-05 18:58:02.806189] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:45.508 18:58:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:45.508 00:07:45.508 real 0m3.270s 00:07:45.508 user 0m5.091s 00:07:45.508 sys 0m0.681s 00:07:45.508 18:58:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:45.508 18:58:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.508 ************************************ 00:07:45.508 END TEST raid_superblock_test 00:07:45.508 ************************************ 00:07:45.768 18:58:03 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 2 read 00:07:45.768 18:58:03 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:45.768 18:58:03 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:45.768 18:58:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:45.768 ************************************ 00:07:45.768 START TEST raid_read_error_test 00:07:45.768 ************************************ 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 2 read 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.JmyfrD3d0Z 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73419 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73419 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 73419 ']' 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:45.768 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:45.768 18:58:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.768 [2024-12-05 18:58:03.190146] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:45.768 [2024-12-05 18:58:03.190736] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73419 ] 00:07:46.028 [2024-12-05 18:58:03.344463] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:46.028 [2024-12-05 18:58:03.368729] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:46.028 [2024-12-05 18:58:03.410619] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:46.028 [2024-12-05 18:58:03.410661] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.599 BaseBdev1_malloc 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.599 true 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.599 [2024-12-05 18:58:04.042067] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:46.599 [2024-12-05 18:58:04.042132] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:46.599 [2024-12-05 18:58:04.042158] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:46.599 [2024-12-05 18:58:04.042173] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:46.599 [2024-12-05 18:58:04.044341] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:46.599 [2024-12-05 18:58:04.044379] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:46.599 BaseBdev1 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.599 BaseBdev2_malloc 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.599 true 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.599 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.599 [2024-12-05 18:58:04.082469] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:46.599 [2024-12-05 18:58:04.082520] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:46.599 [2024-12-05 18:58:04.082536] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:46.599 [2024-12-05 18:58:04.082552] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:46.599 [2024-12-05 18:58:04.084731] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:46.599 [2024-12-05 18:58:04.084767] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:46.600 BaseBdev2 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.600 [2024-12-05 18:58:04.094498] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:46.600 [2024-12-05 18:58:04.096459] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:46.600 [2024-12-05 18:58:04.096718] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:46.600 [2024-12-05 18:58:04.096736] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:46.600 [2024-12-05 18:58:04.097006] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:46.600 [2024-12-05 18:58:04.097159] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:46.600 [2024-12-05 18:58:04.097171] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:46.600 [2024-12-05 18:58:04.097290] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:46.600 "name": "raid_bdev1", 00:07:46.600 "uuid": "1f0d4830-fbf3-4544-ab80-e962a6a6283f", 00:07:46.600 "strip_size_kb": 64, 00:07:46.600 "state": "online", 00:07:46.600 "raid_level": "concat", 00:07:46.600 "superblock": true, 00:07:46.600 "num_base_bdevs": 2, 00:07:46.600 "num_base_bdevs_discovered": 2, 00:07:46.600 "num_base_bdevs_operational": 2, 00:07:46.600 "base_bdevs_list": [ 00:07:46.600 { 00:07:46.600 "name": "BaseBdev1", 00:07:46.600 "uuid": "523eef71-fd0b-587f-8d4f-0bff89a4e414", 00:07:46.600 "is_configured": true, 00:07:46.600 "data_offset": 2048, 00:07:46.600 "data_size": 63488 00:07:46.600 }, 00:07:46.600 { 00:07:46.600 "name": "BaseBdev2", 00:07:46.600 "uuid": "7b179cae-9e10-57fd-a204-c8389a3f0aed", 00:07:46.600 "is_configured": true, 00:07:46.600 "data_offset": 2048, 00:07:46.600 "data_size": 63488 00:07:46.600 } 00:07:46.600 ] 00:07:46.600 }' 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:46.600 18:58:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.169 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:47.169 18:58:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:47.169 [2024-12-05 18:58:04.590043] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:48.117 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:48.118 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:48.118 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.118 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.118 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.118 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:48.118 "name": "raid_bdev1", 00:07:48.118 "uuid": "1f0d4830-fbf3-4544-ab80-e962a6a6283f", 00:07:48.118 "strip_size_kb": 64, 00:07:48.118 "state": "online", 00:07:48.118 "raid_level": "concat", 00:07:48.118 "superblock": true, 00:07:48.118 "num_base_bdevs": 2, 00:07:48.118 "num_base_bdevs_discovered": 2, 00:07:48.118 "num_base_bdevs_operational": 2, 00:07:48.118 "base_bdevs_list": [ 00:07:48.118 { 00:07:48.118 "name": "BaseBdev1", 00:07:48.118 "uuid": "523eef71-fd0b-587f-8d4f-0bff89a4e414", 00:07:48.118 "is_configured": true, 00:07:48.118 "data_offset": 2048, 00:07:48.118 "data_size": 63488 00:07:48.118 }, 00:07:48.118 { 00:07:48.118 "name": "BaseBdev2", 00:07:48.118 "uuid": "7b179cae-9e10-57fd-a204-c8389a3f0aed", 00:07:48.118 "is_configured": true, 00:07:48.118 "data_offset": 2048, 00:07:48.118 "data_size": 63488 00:07:48.118 } 00:07:48.118 ] 00:07:48.118 }' 00:07:48.118 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:48.118 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.392 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:48.392 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:48.392 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.392 [2024-12-05 18:58:05.941692] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:48.392 [2024-12-05 18:58:05.941721] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:48.392 [2024-12-05 18:58:05.944385] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:48.392 [2024-12-05 18:58:05.944486] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:48.392 [2024-12-05 18:58:05.944531] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:48.392 [2024-12-05 18:58:05.944541] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:48.392 { 00:07:48.392 "results": [ 00:07:48.392 { 00:07:48.392 "job": "raid_bdev1", 00:07:48.392 "core_mask": "0x1", 00:07:48.392 "workload": "randrw", 00:07:48.392 "percentage": 50, 00:07:48.392 "status": "finished", 00:07:48.392 "queue_depth": 1, 00:07:48.392 "io_size": 131072, 00:07:48.392 "runtime": 1.352485, 00:07:48.392 "iops": 17346.58794737095, 00:07:48.392 "mibps": 2168.3234934213688, 00:07:48.392 "io_failed": 1, 00:07:48.393 "io_timeout": 0, 00:07:48.393 "avg_latency_us": 79.31757806640042, 00:07:48.393 "min_latency_us": 24.929257641921396, 00:07:48.393 "max_latency_us": 1373.6803493449781 00:07:48.393 } 00:07:48.393 ], 00:07:48.393 "core_count": 1 00:07:48.393 } 00:07:48.393 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:48.393 18:58:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73419 00:07:48.393 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 73419 ']' 00:07:48.393 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 73419 00:07:48.393 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:07:48.652 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:48.652 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73419 00:07:48.652 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:48.652 killing process with pid 73419 00:07:48.653 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:48.653 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73419' 00:07:48.653 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 73419 00:07:48.653 [2024-12-05 18:58:05.994055] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:48.653 18:58:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 73419 00:07:48.653 [2024-12-05 18:58:06.009067] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:48.913 18:58:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:48.914 18:58:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.JmyfrD3d0Z 00:07:48.914 18:58:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:48.914 18:58:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:07:48.914 18:58:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:07:48.914 18:58:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:48.914 18:58:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:48.914 18:58:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:07:48.914 00:07:48.914 real 0m3.128s 00:07:48.914 user 0m3.973s 00:07:48.914 sys 0m0.487s 00:07:48.914 ************************************ 00:07:48.914 END TEST raid_read_error_test 00:07:48.914 ************************************ 00:07:48.914 18:58:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:48.914 18:58:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.914 18:58:06 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 2 write 00:07:48.914 18:58:06 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:48.914 18:58:06 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:48.914 18:58:06 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:48.914 ************************************ 00:07:48.914 START TEST raid_write_error_test 00:07:48.914 ************************************ 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 2 write 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.492wrauQt9 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73549 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73549 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 73549 ']' 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:48.914 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:48.914 18:58:06 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.914 [2024-12-05 18:58:06.385054] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:48.914 [2024-12-05 18:58:06.385187] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73549 ] 00:07:49.174 [2024-12-05 18:58:06.538257] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:49.174 [2024-12-05 18:58:06.562499] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:49.174 [2024-12-05 18:58:06.604274] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:49.174 [2024-12-05 18:58:06.604313] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.743 BaseBdev1_malloc 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.743 true 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.743 [2024-12-05 18:58:07.235793] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:49.743 [2024-12-05 18:58:07.235849] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:49.743 [2024-12-05 18:58:07.235874] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:49.743 [2024-12-05 18:58:07.235889] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:49.743 [2024-12-05 18:58:07.238097] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:49.743 [2024-12-05 18:58:07.238134] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:49.743 BaseBdev1 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.743 BaseBdev2_malloc 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.743 true 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.743 [2024-12-05 18:58:07.276232] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:49.743 [2024-12-05 18:58:07.276279] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:49.743 [2024-12-05 18:58:07.276296] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:49.743 [2024-12-05 18:58:07.276312] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:49.743 [2024-12-05 18:58:07.278489] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:49.743 [2024-12-05 18:58:07.278524] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:49.743 BaseBdev2 00:07:49.743 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.744 [2024-12-05 18:58:07.288260] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:49.744 [2024-12-05 18:58:07.290139] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:49.744 [2024-12-05 18:58:07.290384] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:49.744 [2024-12-05 18:58:07.290401] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:49.744 [2024-12-05 18:58:07.290645] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:07:49.744 [2024-12-05 18:58:07.290811] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:49.744 [2024-12-05 18:58:07.290824] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:49.744 [2024-12-05 18:58:07.290947] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:49.744 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:50.002 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.002 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:50.002 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:50.002 "name": "raid_bdev1", 00:07:50.002 "uuid": "dee1cd2c-91af-41bf-9113-04e56dec43ef", 00:07:50.002 "strip_size_kb": 64, 00:07:50.002 "state": "online", 00:07:50.002 "raid_level": "concat", 00:07:50.002 "superblock": true, 00:07:50.002 "num_base_bdevs": 2, 00:07:50.002 "num_base_bdevs_discovered": 2, 00:07:50.002 "num_base_bdevs_operational": 2, 00:07:50.002 "base_bdevs_list": [ 00:07:50.002 { 00:07:50.002 "name": "BaseBdev1", 00:07:50.002 "uuid": "44258be9-1803-5d11-bae5-db5d3a55001d", 00:07:50.002 "is_configured": true, 00:07:50.002 "data_offset": 2048, 00:07:50.002 "data_size": 63488 00:07:50.002 }, 00:07:50.002 { 00:07:50.002 "name": "BaseBdev2", 00:07:50.002 "uuid": "c3a9c5c0-b67e-5d65-8bfa-5a2e268e95b2", 00:07:50.002 "is_configured": true, 00:07:50.002 "data_offset": 2048, 00:07:50.002 "data_size": 63488 00:07:50.002 } 00:07:50.002 ] 00:07:50.002 }' 00:07:50.002 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:50.002 18:58:07 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.260 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:50.260 18:58:07 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:50.260 [2024-12-05 18:58:07.759879] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:51.195 "name": "raid_bdev1", 00:07:51.195 "uuid": "dee1cd2c-91af-41bf-9113-04e56dec43ef", 00:07:51.195 "strip_size_kb": 64, 00:07:51.195 "state": "online", 00:07:51.195 "raid_level": "concat", 00:07:51.195 "superblock": true, 00:07:51.195 "num_base_bdevs": 2, 00:07:51.195 "num_base_bdevs_discovered": 2, 00:07:51.195 "num_base_bdevs_operational": 2, 00:07:51.195 "base_bdevs_list": [ 00:07:51.195 { 00:07:51.195 "name": "BaseBdev1", 00:07:51.195 "uuid": "44258be9-1803-5d11-bae5-db5d3a55001d", 00:07:51.195 "is_configured": true, 00:07:51.195 "data_offset": 2048, 00:07:51.195 "data_size": 63488 00:07:51.195 }, 00:07:51.195 { 00:07:51.195 "name": "BaseBdev2", 00:07:51.195 "uuid": "c3a9c5c0-b67e-5d65-8bfa-5a2e268e95b2", 00:07:51.195 "is_configured": true, 00:07:51.195 "data_offset": 2048, 00:07:51.195 "data_size": 63488 00:07:51.195 } 00:07:51.195 ] 00:07:51.195 }' 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:51.195 18:58:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.815 18:58:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:51.815 18:58:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:51.815 18:58:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.815 [2024-12-05 18:58:09.119614] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:51.815 [2024-12-05 18:58:09.119726] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:51.815 [2024-12-05 18:58:09.122313] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:51.815 [2024-12-05 18:58:09.122401] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:51.815 [2024-12-05 18:58:09.122456] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:51.815 [2024-12-05 18:58:09.122496] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:51.815 { 00:07:51.815 "results": [ 00:07:51.815 { 00:07:51.815 "job": "raid_bdev1", 00:07:51.815 "core_mask": "0x1", 00:07:51.815 "workload": "randrw", 00:07:51.815 "percentage": 50, 00:07:51.815 "status": "finished", 00:07:51.815 "queue_depth": 1, 00:07:51.815 "io_size": 131072, 00:07:51.815 "runtime": 1.360706, 00:07:51.815 "iops": 17324.094991864516, 00:07:51.815 "mibps": 2165.5118739830646, 00:07:51.815 "io_failed": 1, 00:07:51.815 "io_timeout": 0, 00:07:51.815 "avg_latency_us": 79.52137337300401, 00:07:51.816 "min_latency_us": 24.929257641921396, 00:07:51.816 "max_latency_us": 1373.6803493449781 00:07:51.816 } 00:07:51.816 ], 00:07:51.816 "core_count": 1 00:07:51.816 } 00:07:51.816 18:58:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:51.816 18:58:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73549 00:07:51.816 18:58:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 73549 ']' 00:07:51.816 18:58:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 73549 00:07:51.816 18:58:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:07:51.816 18:58:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:51.816 18:58:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73549 00:07:51.816 killing process with pid 73549 00:07:51.816 18:58:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:51.816 18:58:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:51.816 18:58:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73549' 00:07:51.816 18:58:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 73549 00:07:51.816 [2024-12-05 18:58:09.170559] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:51.816 18:58:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 73549 00:07:51.816 [2024-12-05 18:58:09.185374] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:52.075 18:58:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.492wrauQt9 00:07:52.075 18:58:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:52.075 18:58:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:52.075 18:58:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:07:52.075 18:58:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:07:52.075 18:58:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:52.075 18:58:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:52.075 18:58:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:07:52.075 00:07:52.075 real 0m3.107s 00:07:52.075 user 0m3.940s 00:07:52.075 sys 0m0.485s 00:07:52.075 18:58:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:52.075 ************************************ 00:07:52.075 END TEST raid_write_error_test 00:07:52.075 ************************************ 00:07:52.075 18:58:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.075 18:58:09 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:52.075 18:58:09 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 2 false 00:07:52.075 18:58:09 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:52.075 18:58:09 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:52.075 18:58:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:52.075 ************************************ 00:07:52.075 START TEST raid_state_function_test 00:07:52.075 ************************************ 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 false 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=73675 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:52.075 Process raid pid: 73675 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73675' 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 73675 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 73675 ']' 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:52.075 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:52.075 18:58:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.075 [2024-12-05 18:58:09.558583] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:52.075 [2024-12-05 18:58:09.558832] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:52.335 [2024-12-05 18:58:09.714915] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:52.335 [2024-12-05 18:58:09.739169] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:52.335 [2024-12-05 18:58:09.781134] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:52.335 [2024-12-05 18:58:09.781257] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.904 [2024-12-05 18:58:10.375692] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:52.904 [2024-12-05 18:58:10.375753] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:52.904 [2024-12-05 18:58:10.375763] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:52.904 [2024-12-05 18:58:10.375773] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:52.904 "name": "Existed_Raid", 00:07:52.904 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:52.904 "strip_size_kb": 0, 00:07:52.904 "state": "configuring", 00:07:52.904 "raid_level": "raid1", 00:07:52.904 "superblock": false, 00:07:52.904 "num_base_bdevs": 2, 00:07:52.904 "num_base_bdevs_discovered": 0, 00:07:52.904 "num_base_bdevs_operational": 2, 00:07:52.904 "base_bdevs_list": [ 00:07:52.904 { 00:07:52.904 "name": "BaseBdev1", 00:07:52.904 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:52.904 "is_configured": false, 00:07:52.904 "data_offset": 0, 00:07:52.904 "data_size": 0 00:07:52.904 }, 00:07:52.904 { 00:07:52.904 "name": "BaseBdev2", 00:07:52.904 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:52.904 "is_configured": false, 00:07:52.904 "data_offset": 0, 00:07:52.904 "data_size": 0 00:07:52.904 } 00:07:52.904 ] 00:07:52.904 }' 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:52.904 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.472 [2024-12-05 18:58:10.766908] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:53.472 [2024-12-05 18:58:10.766992] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.472 [2024-12-05 18:58:10.774898] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:53.472 [2024-12-05 18:58:10.774975] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:53.472 [2024-12-05 18:58:10.775002] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:53.472 [2024-12-05 18:58:10.775036] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.472 [2024-12-05 18:58:10.795704] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:53.472 BaseBdev1 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.472 [ 00:07:53.472 { 00:07:53.472 "name": "BaseBdev1", 00:07:53.472 "aliases": [ 00:07:53.472 "c91b7cf6-5bec-4161-a651-3b5878336ffb" 00:07:53.472 ], 00:07:53.472 "product_name": "Malloc disk", 00:07:53.472 "block_size": 512, 00:07:53.472 "num_blocks": 65536, 00:07:53.472 "uuid": "c91b7cf6-5bec-4161-a651-3b5878336ffb", 00:07:53.472 "assigned_rate_limits": { 00:07:53.472 "rw_ios_per_sec": 0, 00:07:53.472 "rw_mbytes_per_sec": 0, 00:07:53.472 "r_mbytes_per_sec": 0, 00:07:53.472 "w_mbytes_per_sec": 0 00:07:53.472 }, 00:07:53.472 "claimed": true, 00:07:53.472 "claim_type": "exclusive_write", 00:07:53.472 "zoned": false, 00:07:53.472 "supported_io_types": { 00:07:53.472 "read": true, 00:07:53.472 "write": true, 00:07:53.472 "unmap": true, 00:07:53.472 "flush": true, 00:07:53.472 "reset": true, 00:07:53.472 "nvme_admin": false, 00:07:53.472 "nvme_io": false, 00:07:53.472 "nvme_io_md": false, 00:07:53.472 "write_zeroes": true, 00:07:53.472 "zcopy": true, 00:07:53.472 "get_zone_info": false, 00:07:53.472 "zone_management": false, 00:07:53.472 "zone_append": false, 00:07:53.472 "compare": false, 00:07:53.472 "compare_and_write": false, 00:07:53.472 "abort": true, 00:07:53.472 "seek_hole": false, 00:07:53.472 "seek_data": false, 00:07:53.472 "copy": true, 00:07:53.472 "nvme_iov_md": false 00:07:53.472 }, 00:07:53.472 "memory_domains": [ 00:07:53.472 { 00:07:53.472 "dma_device_id": "system", 00:07:53.472 "dma_device_type": 1 00:07:53.472 }, 00:07:53.472 { 00:07:53.472 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:53.472 "dma_device_type": 2 00:07:53.472 } 00:07:53.472 ], 00:07:53.472 "driver_specific": {} 00:07:53.472 } 00:07:53.472 ] 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.472 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.473 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:53.473 "name": "Existed_Raid", 00:07:53.473 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.473 "strip_size_kb": 0, 00:07:53.473 "state": "configuring", 00:07:53.473 "raid_level": "raid1", 00:07:53.473 "superblock": false, 00:07:53.473 "num_base_bdevs": 2, 00:07:53.473 "num_base_bdevs_discovered": 1, 00:07:53.473 "num_base_bdevs_operational": 2, 00:07:53.473 "base_bdevs_list": [ 00:07:53.473 { 00:07:53.473 "name": "BaseBdev1", 00:07:53.473 "uuid": "c91b7cf6-5bec-4161-a651-3b5878336ffb", 00:07:53.473 "is_configured": true, 00:07:53.473 "data_offset": 0, 00:07:53.473 "data_size": 65536 00:07:53.473 }, 00:07:53.473 { 00:07:53.473 "name": "BaseBdev2", 00:07:53.473 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.473 "is_configured": false, 00:07:53.473 "data_offset": 0, 00:07:53.473 "data_size": 0 00:07:53.473 } 00:07:53.473 ] 00:07:53.473 }' 00:07:53.473 18:58:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:53.473 18:58:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.731 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:53.731 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.731 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.731 [2024-12-05 18:58:11.262894] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:53.731 [2024-12-05 18:58:11.262979] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:53.731 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.731 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:53.731 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.731 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.731 [2024-12-05 18:58:11.274897] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:53.731 [2024-12-05 18:58:11.276794] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:53.732 [2024-12-05 18:58:11.276886] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:53.732 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.990 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:53.990 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:53.990 "name": "Existed_Raid", 00:07:53.990 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.990 "strip_size_kb": 0, 00:07:53.990 "state": "configuring", 00:07:53.990 "raid_level": "raid1", 00:07:53.990 "superblock": false, 00:07:53.990 "num_base_bdevs": 2, 00:07:53.990 "num_base_bdevs_discovered": 1, 00:07:53.990 "num_base_bdevs_operational": 2, 00:07:53.990 "base_bdevs_list": [ 00:07:53.990 { 00:07:53.990 "name": "BaseBdev1", 00:07:53.990 "uuid": "c91b7cf6-5bec-4161-a651-3b5878336ffb", 00:07:53.990 "is_configured": true, 00:07:53.990 "data_offset": 0, 00:07:53.990 "data_size": 65536 00:07:53.990 }, 00:07:53.990 { 00:07:53.990 "name": "BaseBdev2", 00:07:53.990 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.990 "is_configured": false, 00:07:53.990 "data_offset": 0, 00:07:53.990 "data_size": 0 00:07:53.990 } 00:07:53.990 ] 00:07:53.990 }' 00:07:53.990 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:53.990 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.249 [2024-12-05 18:58:11.657243] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:54.249 [2024-12-05 18:58:11.657376] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:54.249 [2024-12-05 18:58:11.657401] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:07:54.249 [2024-12-05 18:58:11.657718] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:54.249 [2024-12-05 18:58:11.657896] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:54.249 [2024-12-05 18:58:11.657943] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:54.249 [2024-12-05 18:58:11.658179] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:54.249 BaseBdev2 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.249 [ 00:07:54.249 { 00:07:54.249 "name": "BaseBdev2", 00:07:54.249 "aliases": [ 00:07:54.249 "638a456c-09c3-49d6-b7ab-f1578d8da96c" 00:07:54.249 ], 00:07:54.249 "product_name": "Malloc disk", 00:07:54.249 "block_size": 512, 00:07:54.249 "num_blocks": 65536, 00:07:54.249 "uuid": "638a456c-09c3-49d6-b7ab-f1578d8da96c", 00:07:54.249 "assigned_rate_limits": { 00:07:54.249 "rw_ios_per_sec": 0, 00:07:54.249 "rw_mbytes_per_sec": 0, 00:07:54.249 "r_mbytes_per_sec": 0, 00:07:54.249 "w_mbytes_per_sec": 0 00:07:54.249 }, 00:07:54.249 "claimed": true, 00:07:54.249 "claim_type": "exclusive_write", 00:07:54.249 "zoned": false, 00:07:54.249 "supported_io_types": { 00:07:54.249 "read": true, 00:07:54.249 "write": true, 00:07:54.249 "unmap": true, 00:07:54.249 "flush": true, 00:07:54.249 "reset": true, 00:07:54.249 "nvme_admin": false, 00:07:54.249 "nvme_io": false, 00:07:54.249 "nvme_io_md": false, 00:07:54.249 "write_zeroes": true, 00:07:54.249 "zcopy": true, 00:07:54.249 "get_zone_info": false, 00:07:54.249 "zone_management": false, 00:07:54.249 "zone_append": false, 00:07:54.249 "compare": false, 00:07:54.249 "compare_and_write": false, 00:07:54.249 "abort": true, 00:07:54.249 "seek_hole": false, 00:07:54.249 "seek_data": false, 00:07:54.249 "copy": true, 00:07:54.249 "nvme_iov_md": false 00:07:54.249 }, 00:07:54.249 "memory_domains": [ 00:07:54.249 { 00:07:54.249 "dma_device_id": "system", 00:07:54.249 "dma_device_type": 1 00:07:54.249 }, 00:07:54.249 { 00:07:54.249 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:54.249 "dma_device_type": 2 00:07:54.249 } 00:07:54.249 ], 00:07:54.249 "driver_specific": {} 00:07:54.249 } 00:07:54.249 ] 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:54.249 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:54.250 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:54.250 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.250 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:54.250 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.250 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.250 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.250 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:54.250 "name": "Existed_Raid", 00:07:54.250 "uuid": "74d6efe2-23d7-4ce4-8080-863aa62f5b84", 00:07:54.250 "strip_size_kb": 0, 00:07:54.250 "state": "online", 00:07:54.250 "raid_level": "raid1", 00:07:54.250 "superblock": false, 00:07:54.250 "num_base_bdevs": 2, 00:07:54.250 "num_base_bdevs_discovered": 2, 00:07:54.250 "num_base_bdevs_operational": 2, 00:07:54.250 "base_bdevs_list": [ 00:07:54.250 { 00:07:54.250 "name": "BaseBdev1", 00:07:54.250 "uuid": "c91b7cf6-5bec-4161-a651-3b5878336ffb", 00:07:54.250 "is_configured": true, 00:07:54.250 "data_offset": 0, 00:07:54.250 "data_size": 65536 00:07:54.250 }, 00:07:54.250 { 00:07:54.250 "name": "BaseBdev2", 00:07:54.250 "uuid": "638a456c-09c3-49d6-b7ab-f1578d8da96c", 00:07:54.250 "is_configured": true, 00:07:54.250 "data_offset": 0, 00:07:54.250 "data_size": 65536 00:07:54.250 } 00:07:54.250 ] 00:07:54.250 }' 00:07:54.250 18:58:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:54.250 18:58:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.818 [2024-12-05 18:58:12.124783] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:54.818 "name": "Existed_Raid", 00:07:54.818 "aliases": [ 00:07:54.818 "74d6efe2-23d7-4ce4-8080-863aa62f5b84" 00:07:54.818 ], 00:07:54.818 "product_name": "Raid Volume", 00:07:54.818 "block_size": 512, 00:07:54.818 "num_blocks": 65536, 00:07:54.818 "uuid": "74d6efe2-23d7-4ce4-8080-863aa62f5b84", 00:07:54.818 "assigned_rate_limits": { 00:07:54.818 "rw_ios_per_sec": 0, 00:07:54.818 "rw_mbytes_per_sec": 0, 00:07:54.818 "r_mbytes_per_sec": 0, 00:07:54.818 "w_mbytes_per_sec": 0 00:07:54.818 }, 00:07:54.818 "claimed": false, 00:07:54.818 "zoned": false, 00:07:54.818 "supported_io_types": { 00:07:54.818 "read": true, 00:07:54.818 "write": true, 00:07:54.818 "unmap": false, 00:07:54.818 "flush": false, 00:07:54.818 "reset": true, 00:07:54.818 "nvme_admin": false, 00:07:54.818 "nvme_io": false, 00:07:54.818 "nvme_io_md": false, 00:07:54.818 "write_zeroes": true, 00:07:54.818 "zcopy": false, 00:07:54.818 "get_zone_info": false, 00:07:54.818 "zone_management": false, 00:07:54.818 "zone_append": false, 00:07:54.818 "compare": false, 00:07:54.818 "compare_and_write": false, 00:07:54.818 "abort": false, 00:07:54.818 "seek_hole": false, 00:07:54.818 "seek_data": false, 00:07:54.818 "copy": false, 00:07:54.818 "nvme_iov_md": false 00:07:54.818 }, 00:07:54.818 "memory_domains": [ 00:07:54.818 { 00:07:54.818 "dma_device_id": "system", 00:07:54.818 "dma_device_type": 1 00:07:54.818 }, 00:07:54.818 { 00:07:54.818 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:54.818 "dma_device_type": 2 00:07:54.818 }, 00:07:54.818 { 00:07:54.818 "dma_device_id": "system", 00:07:54.818 "dma_device_type": 1 00:07:54.818 }, 00:07:54.818 { 00:07:54.818 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:54.818 "dma_device_type": 2 00:07:54.818 } 00:07:54.818 ], 00:07:54.818 "driver_specific": { 00:07:54.818 "raid": { 00:07:54.818 "uuid": "74d6efe2-23d7-4ce4-8080-863aa62f5b84", 00:07:54.818 "strip_size_kb": 0, 00:07:54.818 "state": "online", 00:07:54.818 "raid_level": "raid1", 00:07:54.818 "superblock": false, 00:07:54.818 "num_base_bdevs": 2, 00:07:54.818 "num_base_bdevs_discovered": 2, 00:07:54.818 "num_base_bdevs_operational": 2, 00:07:54.818 "base_bdevs_list": [ 00:07:54.818 { 00:07:54.818 "name": "BaseBdev1", 00:07:54.818 "uuid": "c91b7cf6-5bec-4161-a651-3b5878336ffb", 00:07:54.818 "is_configured": true, 00:07:54.818 "data_offset": 0, 00:07:54.818 "data_size": 65536 00:07:54.818 }, 00:07:54.818 { 00:07:54.818 "name": "BaseBdev2", 00:07:54.818 "uuid": "638a456c-09c3-49d6-b7ab-f1578d8da96c", 00:07:54.818 "is_configured": true, 00:07:54.818 "data_offset": 0, 00:07:54.818 "data_size": 65536 00:07:54.818 } 00:07:54.818 ] 00:07:54.818 } 00:07:54.818 } 00:07:54.818 }' 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:54.818 BaseBdev2' 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.818 [2024-12-05 18:58:12.348150] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:54.818 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.078 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.078 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:55.078 "name": "Existed_Raid", 00:07:55.078 "uuid": "74d6efe2-23d7-4ce4-8080-863aa62f5b84", 00:07:55.078 "strip_size_kb": 0, 00:07:55.078 "state": "online", 00:07:55.078 "raid_level": "raid1", 00:07:55.078 "superblock": false, 00:07:55.078 "num_base_bdevs": 2, 00:07:55.078 "num_base_bdevs_discovered": 1, 00:07:55.078 "num_base_bdevs_operational": 1, 00:07:55.078 "base_bdevs_list": [ 00:07:55.078 { 00:07:55.078 "name": null, 00:07:55.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.078 "is_configured": false, 00:07:55.078 "data_offset": 0, 00:07:55.078 "data_size": 65536 00:07:55.078 }, 00:07:55.078 { 00:07:55.078 "name": "BaseBdev2", 00:07:55.078 "uuid": "638a456c-09c3-49d6-b7ab-f1578d8da96c", 00:07:55.078 "is_configured": true, 00:07:55.078 "data_offset": 0, 00:07:55.078 "data_size": 65536 00:07:55.078 } 00:07:55.078 ] 00:07:55.078 }' 00:07:55.078 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:55.078 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.337 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.338 [2024-12-05 18:58:12.846529] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:55.338 [2024-12-05 18:58:12.846674] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:55.338 [2024-12-05 18:58:12.857921] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:55.338 [2024-12-05 18:58:12.858030] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:55.338 [2024-12-05 18:58:12.858072] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.338 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:55.598 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:55.598 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:55.598 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:55.598 18:58:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 73675 00:07:55.598 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 73675 ']' 00:07:55.598 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 73675 00:07:55.598 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:07:55.598 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:55.598 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73675 00:07:55.598 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:55.598 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:55.598 killing process with pid 73675 00:07:55.598 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73675' 00:07:55.598 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 73675 00:07:55.598 [2024-12-05 18:58:12.945297] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:55.598 18:58:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 73675 00:07:55.598 [2024-12-05 18:58:12.946340] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:55.858 00:07:55.858 real 0m3.692s 00:07:55.858 user 0m5.827s 00:07:55.858 sys 0m0.731s 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.858 ************************************ 00:07:55.858 END TEST raid_state_function_test 00:07:55.858 ************************************ 00:07:55.858 18:58:13 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 2 true 00:07:55.858 18:58:13 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:07:55.858 18:58:13 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:55.858 18:58:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:55.858 ************************************ 00:07:55.858 START TEST raid_state_function_test_sb 00:07:55.858 ************************************ 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:55.858 Process raid pid: 73913 00:07:55.858 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=73913 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73913' 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 73913 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 73913 ']' 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:55.858 18:58:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:55.858 [2024-12-05 18:58:13.322779] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:55.858 [2024-12-05 18:58:13.322894] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:56.118 [2024-12-05 18:58:13.476516] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:56.118 [2024-12-05 18:58:13.500708] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:56.118 [2024-12-05 18:58:13.542437] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:56.118 [2024-12-05 18:58:13.542478] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.689 [2024-12-05 18:58:14.140976] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:56.689 [2024-12-05 18:58:14.141038] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:56.689 [2024-12-05 18:58:14.141048] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:56.689 [2024-12-05 18:58:14.141059] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:56.689 "name": "Existed_Raid", 00:07:56.689 "uuid": "4555f0fc-634f-4fec-9158-0c8fa1945a54", 00:07:56.689 "strip_size_kb": 0, 00:07:56.689 "state": "configuring", 00:07:56.689 "raid_level": "raid1", 00:07:56.689 "superblock": true, 00:07:56.689 "num_base_bdevs": 2, 00:07:56.689 "num_base_bdevs_discovered": 0, 00:07:56.689 "num_base_bdevs_operational": 2, 00:07:56.689 "base_bdevs_list": [ 00:07:56.689 { 00:07:56.689 "name": "BaseBdev1", 00:07:56.689 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:56.689 "is_configured": false, 00:07:56.689 "data_offset": 0, 00:07:56.689 "data_size": 0 00:07:56.689 }, 00:07:56.689 { 00:07:56.689 "name": "BaseBdev2", 00:07:56.689 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:56.689 "is_configured": false, 00:07:56.689 "data_offset": 0, 00:07:56.689 "data_size": 0 00:07:56.689 } 00:07:56.689 ] 00:07:56.689 }' 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:56.689 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.260 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:57.260 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.260 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.260 [2024-12-05 18:58:14.568128] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:57.260 [2024-12-05 18:58:14.568224] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:57.260 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.260 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:57.260 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.260 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.260 [2024-12-05 18:58:14.580123] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:57.260 [2024-12-05 18:58:14.580217] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:57.260 [2024-12-05 18:58:14.580243] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:57.260 [2024-12-05 18:58:14.580276] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:57.260 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.260 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.261 [2024-12-05 18:58:14.600905] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:57.261 BaseBdev1 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.261 [ 00:07:57.261 { 00:07:57.261 "name": "BaseBdev1", 00:07:57.261 "aliases": [ 00:07:57.261 "5c263cb7-b20c-44a0-9a63-a1c6dba5e8cc" 00:07:57.261 ], 00:07:57.261 "product_name": "Malloc disk", 00:07:57.261 "block_size": 512, 00:07:57.261 "num_blocks": 65536, 00:07:57.261 "uuid": "5c263cb7-b20c-44a0-9a63-a1c6dba5e8cc", 00:07:57.261 "assigned_rate_limits": { 00:07:57.261 "rw_ios_per_sec": 0, 00:07:57.261 "rw_mbytes_per_sec": 0, 00:07:57.261 "r_mbytes_per_sec": 0, 00:07:57.261 "w_mbytes_per_sec": 0 00:07:57.261 }, 00:07:57.261 "claimed": true, 00:07:57.261 "claim_type": "exclusive_write", 00:07:57.261 "zoned": false, 00:07:57.261 "supported_io_types": { 00:07:57.261 "read": true, 00:07:57.261 "write": true, 00:07:57.261 "unmap": true, 00:07:57.261 "flush": true, 00:07:57.261 "reset": true, 00:07:57.261 "nvme_admin": false, 00:07:57.261 "nvme_io": false, 00:07:57.261 "nvme_io_md": false, 00:07:57.261 "write_zeroes": true, 00:07:57.261 "zcopy": true, 00:07:57.261 "get_zone_info": false, 00:07:57.261 "zone_management": false, 00:07:57.261 "zone_append": false, 00:07:57.261 "compare": false, 00:07:57.261 "compare_and_write": false, 00:07:57.261 "abort": true, 00:07:57.261 "seek_hole": false, 00:07:57.261 "seek_data": false, 00:07:57.261 "copy": true, 00:07:57.261 "nvme_iov_md": false 00:07:57.261 }, 00:07:57.261 "memory_domains": [ 00:07:57.261 { 00:07:57.261 "dma_device_id": "system", 00:07:57.261 "dma_device_type": 1 00:07:57.261 }, 00:07:57.261 { 00:07:57.261 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:57.261 "dma_device_type": 2 00:07:57.261 } 00:07:57.261 ], 00:07:57.261 "driver_specific": {} 00:07:57.261 } 00:07:57.261 ] 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:57.261 "name": "Existed_Raid", 00:07:57.261 "uuid": "e36a905b-d112-4389-898a-6d943d49677d", 00:07:57.261 "strip_size_kb": 0, 00:07:57.261 "state": "configuring", 00:07:57.261 "raid_level": "raid1", 00:07:57.261 "superblock": true, 00:07:57.261 "num_base_bdevs": 2, 00:07:57.261 "num_base_bdevs_discovered": 1, 00:07:57.261 "num_base_bdevs_operational": 2, 00:07:57.261 "base_bdevs_list": [ 00:07:57.261 { 00:07:57.261 "name": "BaseBdev1", 00:07:57.261 "uuid": "5c263cb7-b20c-44a0-9a63-a1c6dba5e8cc", 00:07:57.261 "is_configured": true, 00:07:57.261 "data_offset": 2048, 00:07:57.261 "data_size": 63488 00:07:57.261 }, 00:07:57.261 { 00:07:57.261 "name": "BaseBdev2", 00:07:57.261 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.261 "is_configured": false, 00:07:57.261 "data_offset": 0, 00:07:57.261 "data_size": 0 00:07:57.261 } 00:07:57.261 ] 00:07:57.261 }' 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:57.261 18:58:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.831 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:57.831 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.831 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.831 [2024-12-05 18:58:15.084113] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:57.831 [2024-12-05 18:58:15.084202] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.832 [2024-12-05 18:58:15.096126] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:57.832 [2024-12-05 18:58:15.098035] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:57.832 [2024-12-05 18:58:15.098120] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:57.832 "name": "Existed_Raid", 00:07:57.832 "uuid": "c5995d9f-ffcb-445e-985f-f1ec920465bc", 00:07:57.832 "strip_size_kb": 0, 00:07:57.832 "state": "configuring", 00:07:57.832 "raid_level": "raid1", 00:07:57.832 "superblock": true, 00:07:57.832 "num_base_bdevs": 2, 00:07:57.832 "num_base_bdevs_discovered": 1, 00:07:57.832 "num_base_bdevs_operational": 2, 00:07:57.832 "base_bdevs_list": [ 00:07:57.832 { 00:07:57.832 "name": "BaseBdev1", 00:07:57.832 "uuid": "5c263cb7-b20c-44a0-9a63-a1c6dba5e8cc", 00:07:57.832 "is_configured": true, 00:07:57.832 "data_offset": 2048, 00:07:57.832 "data_size": 63488 00:07:57.832 }, 00:07:57.832 { 00:07:57.832 "name": "BaseBdev2", 00:07:57.832 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.832 "is_configured": false, 00:07:57.832 "data_offset": 0, 00:07:57.832 "data_size": 0 00:07:57.832 } 00:07:57.832 ] 00:07:57.832 }' 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:57.832 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.091 [2024-12-05 18:58:15.570202] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:58.091 [2024-12-05 18:58:15.570495] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:58.091 [2024-12-05 18:58:15.570544] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:58.091 BaseBdev2 00:07:58.091 [2024-12-05 18:58:15.570870] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:58.091 [2024-12-05 18:58:15.571068] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:58.091 [2024-12-05 18:58:15.571117] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:58.091 [2024-12-05 18:58:15.571283] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.091 [ 00:07:58.091 { 00:07:58.091 "name": "BaseBdev2", 00:07:58.091 "aliases": [ 00:07:58.091 "bf7a81a0-c5a8-41bd-8965-1be6073da2d6" 00:07:58.091 ], 00:07:58.091 "product_name": "Malloc disk", 00:07:58.091 "block_size": 512, 00:07:58.091 "num_blocks": 65536, 00:07:58.091 "uuid": "bf7a81a0-c5a8-41bd-8965-1be6073da2d6", 00:07:58.091 "assigned_rate_limits": { 00:07:58.091 "rw_ios_per_sec": 0, 00:07:58.091 "rw_mbytes_per_sec": 0, 00:07:58.091 "r_mbytes_per_sec": 0, 00:07:58.091 "w_mbytes_per_sec": 0 00:07:58.091 }, 00:07:58.091 "claimed": true, 00:07:58.091 "claim_type": "exclusive_write", 00:07:58.091 "zoned": false, 00:07:58.091 "supported_io_types": { 00:07:58.091 "read": true, 00:07:58.091 "write": true, 00:07:58.091 "unmap": true, 00:07:58.091 "flush": true, 00:07:58.091 "reset": true, 00:07:58.091 "nvme_admin": false, 00:07:58.091 "nvme_io": false, 00:07:58.091 "nvme_io_md": false, 00:07:58.091 "write_zeroes": true, 00:07:58.091 "zcopy": true, 00:07:58.091 "get_zone_info": false, 00:07:58.091 "zone_management": false, 00:07:58.091 "zone_append": false, 00:07:58.091 "compare": false, 00:07:58.091 "compare_and_write": false, 00:07:58.091 "abort": true, 00:07:58.091 "seek_hole": false, 00:07:58.091 "seek_data": false, 00:07:58.091 "copy": true, 00:07:58.091 "nvme_iov_md": false 00:07:58.091 }, 00:07:58.091 "memory_domains": [ 00:07:58.091 { 00:07:58.091 "dma_device_id": "system", 00:07:58.091 "dma_device_type": 1 00:07:58.091 }, 00:07:58.091 { 00:07:58.091 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:58.091 "dma_device_type": 2 00:07:58.091 } 00:07:58.091 ], 00:07:58.091 "driver_specific": {} 00:07:58.091 } 00:07:58.091 ] 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.091 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.351 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:58.351 "name": "Existed_Raid", 00:07:58.351 "uuid": "c5995d9f-ffcb-445e-985f-f1ec920465bc", 00:07:58.351 "strip_size_kb": 0, 00:07:58.351 "state": "online", 00:07:58.351 "raid_level": "raid1", 00:07:58.351 "superblock": true, 00:07:58.351 "num_base_bdevs": 2, 00:07:58.351 "num_base_bdevs_discovered": 2, 00:07:58.351 "num_base_bdevs_operational": 2, 00:07:58.351 "base_bdevs_list": [ 00:07:58.351 { 00:07:58.351 "name": "BaseBdev1", 00:07:58.351 "uuid": "5c263cb7-b20c-44a0-9a63-a1c6dba5e8cc", 00:07:58.351 "is_configured": true, 00:07:58.351 "data_offset": 2048, 00:07:58.351 "data_size": 63488 00:07:58.351 }, 00:07:58.351 { 00:07:58.351 "name": "BaseBdev2", 00:07:58.351 "uuid": "bf7a81a0-c5a8-41bd-8965-1be6073da2d6", 00:07:58.351 "is_configured": true, 00:07:58.351 "data_offset": 2048, 00:07:58.351 "data_size": 63488 00:07:58.351 } 00:07:58.351 ] 00:07:58.351 }' 00:07:58.351 18:58:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:58.351 18:58:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.611 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:58.611 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:58.611 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:58.611 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:58.611 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:58.611 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:58.611 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:58.611 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:58.611 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.611 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.611 [2024-12-05 18:58:16.049663] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:58.611 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.611 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:58.611 "name": "Existed_Raid", 00:07:58.611 "aliases": [ 00:07:58.611 "c5995d9f-ffcb-445e-985f-f1ec920465bc" 00:07:58.611 ], 00:07:58.611 "product_name": "Raid Volume", 00:07:58.611 "block_size": 512, 00:07:58.611 "num_blocks": 63488, 00:07:58.611 "uuid": "c5995d9f-ffcb-445e-985f-f1ec920465bc", 00:07:58.611 "assigned_rate_limits": { 00:07:58.611 "rw_ios_per_sec": 0, 00:07:58.611 "rw_mbytes_per_sec": 0, 00:07:58.611 "r_mbytes_per_sec": 0, 00:07:58.611 "w_mbytes_per_sec": 0 00:07:58.611 }, 00:07:58.611 "claimed": false, 00:07:58.611 "zoned": false, 00:07:58.611 "supported_io_types": { 00:07:58.611 "read": true, 00:07:58.611 "write": true, 00:07:58.611 "unmap": false, 00:07:58.611 "flush": false, 00:07:58.611 "reset": true, 00:07:58.611 "nvme_admin": false, 00:07:58.611 "nvme_io": false, 00:07:58.611 "nvme_io_md": false, 00:07:58.611 "write_zeroes": true, 00:07:58.611 "zcopy": false, 00:07:58.611 "get_zone_info": false, 00:07:58.611 "zone_management": false, 00:07:58.611 "zone_append": false, 00:07:58.611 "compare": false, 00:07:58.611 "compare_and_write": false, 00:07:58.611 "abort": false, 00:07:58.611 "seek_hole": false, 00:07:58.611 "seek_data": false, 00:07:58.611 "copy": false, 00:07:58.611 "nvme_iov_md": false 00:07:58.611 }, 00:07:58.611 "memory_domains": [ 00:07:58.611 { 00:07:58.611 "dma_device_id": "system", 00:07:58.611 "dma_device_type": 1 00:07:58.611 }, 00:07:58.611 { 00:07:58.611 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:58.611 "dma_device_type": 2 00:07:58.611 }, 00:07:58.611 { 00:07:58.611 "dma_device_id": "system", 00:07:58.611 "dma_device_type": 1 00:07:58.611 }, 00:07:58.611 { 00:07:58.611 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:58.611 "dma_device_type": 2 00:07:58.611 } 00:07:58.611 ], 00:07:58.611 "driver_specific": { 00:07:58.611 "raid": { 00:07:58.611 "uuid": "c5995d9f-ffcb-445e-985f-f1ec920465bc", 00:07:58.612 "strip_size_kb": 0, 00:07:58.612 "state": "online", 00:07:58.612 "raid_level": "raid1", 00:07:58.612 "superblock": true, 00:07:58.612 "num_base_bdevs": 2, 00:07:58.612 "num_base_bdevs_discovered": 2, 00:07:58.612 "num_base_bdevs_operational": 2, 00:07:58.612 "base_bdevs_list": [ 00:07:58.612 { 00:07:58.612 "name": "BaseBdev1", 00:07:58.612 "uuid": "5c263cb7-b20c-44a0-9a63-a1c6dba5e8cc", 00:07:58.612 "is_configured": true, 00:07:58.612 "data_offset": 2048, 00:07:58.612 "data_size": 63488 00:07:58.612 }, 00:07:58.612 { 00:07:58.612 "name": "BaseBdev2", 00:07:58.612 "uuid": "bf7a81a0-c5a8-41bd-8965-1be6073da2d6", 00:07:58.612 "is_configured": true, 00:07:58.612 "data_offset": 2048, 00:07:58.612 "data_size": 63488 00:07:58.612 } 00:07:58.612 ] 00:07:58.612 } 00:07:58.612 } 00:07:58.612 }' 00:07:58.612 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:58.612 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:58.612 BaseBdev2' 00:07:58.612 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:58.612 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:58.612 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:58.612 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:58.612 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.612 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.612 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.872 [2024-12-05 18:58:16.265104] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:58.872 "name": "Existed_Raid", 00:07:58.872 "uuid": "c5995d9f-ffcb-445e-985f-f1ec920465bc", 00:07:58.872 "strip_size_kb": 0, 00:07:58.872 "state": "online", 00:07:58.872 "raid_level": "raid1", 00:07:58.872 "superblock": true, 00:07:58.872 "num_base_bdevs": 2, 00:07:58.872 "num_base_bdevs_discovered": 1, 00:07:58.872 "num_base_bdevs_operational": 1, 00:07:58.872 "base_bdevs_list": [ 00:07:58.872 { 00:07:58.872 "name": null, 00:07:58.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:58.872 "is_configured": false, 00:07:58.872 "data_offset": 0, 00:07:58.872 "data_size": 63488 00:07:58.872 }, 00:07:58.872 { 00:07:58.872 "name": "BaseBdev2", 00:07:58.872 "uuid": "bf7a81a0-c5a8-41bd-8965-1be6073da2d6", 00:07:58.872 "is_configured": true, 00:07:58.872 "data_offset": 2048, 00:07:58.872 "data_size": 63488 00:07:58.872 } 00:07:58.872 ] 00:07:58.872 }' 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:58.872 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.443 [2024-12-05 18:58:16.727409] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:59.443 [2024-12-05 18:58:16.727500] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:59.443 [2024-12-05 18:58:16.738977] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:59.443 [2024-12-05 18:58:16.739027] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:59.443 [2024-12-05 18:58:16.739039] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 73913 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 73913 ']' 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 73913 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 73913 00:07:59.443 killing process with pid 73913 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 73913' 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 73913 00:07:59.443 [2024-12-05 18:58:16.822057] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:59.443 18:58:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 73913 00:07:59.443 [2024-12-05 18:58:16.823006] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:59.705 18:58:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:59.705 00:07:59.705 real 0m3.804s 00:07:59.705 user 0m6.040s 00:07:59.705 sys 0m0.728s 00:07:59.705 18:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:07:59.705 18:58:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:59.705 ************************************ 00:07:59.705 END TEST raid_state_function_test_sb 00:07:59.705 ************************************ 00:07:59.705 18:58:17 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 2 00:07:59.705 18:58:17 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:07:59.705 18:58:17 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:07:59.705 18:58:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:59.705 ************************************ 00:07:59.705 START TEST raid_superblock_test 00:07:59.705 ************************************ 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=74148 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 74148 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 74148 ']' 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:07:59.705 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:07:59.705 18:58:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.705 [2024-12-05 18:58:17.192216] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:07:59.705 [2024-12-05 18:58:17.192355] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74148 ] 00:07:59.970 [2024-12-05 18:58:17.345242] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:59.970 [2024-12-05 18:58:17.369853] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:59.970 [2024-12-05 18:58:17.411668] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:59.970 [2024-12-05 18:58:17.411724] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.540 malloc1 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.540 [2024-12-05 18:58:18.046869] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:00.540 [2024-12-05 18:58:18.047004] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:00.540 [2024-12-05 18:58:18.047044] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:08:00.540 [2024-12-05 18:58:18.047110] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:00.540 [2024-12-05 18:58:18.049180] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:00.540 [2024-12-05 18:58:18.049269] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:00.540 pt1 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.540 malloc2 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.540 [2024-12-05 18:58:18.079344] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:00.540 [2024-12-05 18:58:18.079401] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:00.540 [2024-12-05 18:58:18.079419] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:00.540 [2024-12-05 18:58:18.079429] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:00.540 [2024-12-05 18:58:18.081577] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:00.540 [2024-12-05 18:58:18.081612] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:00.540 pt2 00:08:00.540 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.541 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:00.541 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:00.541 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:08:00.541 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.541 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.541 [2024-12-05 18:58:18.091347] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:00.541 [2024-12-05 18:58:18.093174] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:00.541 [2024-12-05 18:58:18.093370] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:00.541 [2024-12-05 18:58:18.093389] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:00.541 [2024-12-05 18:58:18.093631] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:00.541 [2024-12-05 18:58:18.093784] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:00.541 [2024-12-05 18:58:18.093795] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:08:00.541 [2024-12-05 18:58:18.093913] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:00.801 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.801 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:00.801 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:00.801 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:00.801 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:00.801 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:00.801 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:00.801 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:00.801 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:00.802 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:00.802 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:00.802 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.802 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:00.802 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:00.802 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.802 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:00.802 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:00.802 "name": "raid_bdev1", 00:08:00.802 "uuid": "5a793038-9d08-4e17-aabd-0542827c168b", 00:08:00.802 "strip_size_kb": 0, 00:08:00.802 "state": "online", 00:08:00.802 "raid_level": "raid1", 00:08:00.802 "superblock": true, 00:08:00.802 "num_base_bdevs": 2, 00:08:00.802 "num_base_bdevs_discovered": 2, 00:08:00.802 "num_base_bdevs_operational": 2, 00:08:00.802 "base_bdevs_list": [ 00:08:00.802 { 00:08:00.802 "name": "pt1", 00:08:00.802 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:00.802 "is_configured": true, 00:08:00.802 "data_offset": 2048, 00:08:00.802 "data_size": 63488 00:08:00.802 }, 00:08:00.802 { 00:08:00.802 "name": "pt2", 00:08:00.802 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:00.802 "is_configured": true, 00:08:00.802 "data_offset": 2048, 00:08:00.802 "data_size": 63488 00:08:00.802 } 00:08:00.802 ] 00:08:00.802 }' 00:08:00.802 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:00.802 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.063 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:01.063 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:01.063 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:01.063 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:01.063 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:01.063 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:01.063 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:01.063 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:01.063 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.063 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.063 [2024-12-05 18:58:18.526892] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:01.063 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.063 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:01.063 "name": "raid_bdev1", 00:08:01.063 "aliases": [ 00:08:01.063 "5a793038-9d08-4e17-aabd-0542827c168b" 00:08:01.063 ], 00:08:01.063 "product_name": "Raid Volume", 00:08:01.063 "block_size": 512, 00:08:01.063 "num_blocks": 63488, 00:08:01.063 "uuid": "5a793038-9d08-4e17-aabd-0542827c168b", 00:08:01.063 "assigned_rate_limits": { 00:08:01.063 "rw_ios_per_sec": 0, 00:08:01.063 "rw_mbytes_per_sec": 0, 00:08:01.063 "r_mbytes_per_sec": 0, 00:08:01.063 "w_mbytes_per_sec": 0 00:08:01.063 }, 00:08:01.063 "claimed": false, 00:08:01.063 "zoned": false, 00:08:01.063 "supported_io_types": { 00:08:01.063 "read": true, 00:08:01.063 "write": true, 00:08:01.063 "unmap": false, 00:08:01.063 "flush": false, 00:08:01.063 "reset": true, 00:08:01.063 "nvme_admin": false, 00:08:01.063 "nvme_io": false, 00:08:01.063 "nvme_io_md": false, 00:08:01.063 "write_zeroes": true, 00:08:01.063 "zcopy": false, 00:08:01.063 "get_zone_info": false, 00:08:01.063 "zone_management": false, 00:08:01.063 "zone_append": false, 00:08:01.063 "compare": false, 00:08:01.063 "compare_and_write": false, 00:08:01.063 "abort": false, 00:08:01.063 "seek_hole": false, 00:08:01.063 "seek_data": false, 00:08:01.063 "copy": false, 00:08:01.063 "nvme_iov_md": false 00:08:01.063 }, 00:08:01.063 "memory_domains": [ 00:08:01.063 { 00:08:01.063 "dma_device_id": "system", 00:08:01.063 "dma_device_type": 1 00:08:01.063 }, 00:08:01.063 { 00:08:01.063 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:01.063 "dma_device_type": 2 00:08:01.063 }, 00:08:01.063 { 00:08:01.063 "dma_device_id": "system", 00:08:01.063 "dma_device_type": 1 00:08:01.063 }, 00:08:01.063 { 00:08:01.063 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:01.063 "dma_device_type": 2 00:08:01.063 } 00:08:01.063 ], 00:08:01.063 "driver_specific": { 00:08:01.063 "raid": { 00:08:01.063 "uuid": "5a793038-9d08-4e17-aabd-0542827c168b", 00:08:01.063 "strip_size_kb": 0, 00:08:01.063 "state": "online", 00:08:01.063 "raid_level": "raid1", 00:08:01.063 "superblock": true, 00:08:01.063 "num_base_bdevs": 2, 00:08:01.063 "num_base_bdevs_discovered": 2, 00:08:01.063 "num_base_bdevs_operational": 2, 00:08:01.063 "base_bdevs_list": [ 00:08:01.063 { 00:08:01.063 "name": "pt1", 00:08:01.063 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:01.063 "is_configured": true, 00:08:01.063 "data_offset": 2048, 00:08:01.063 "data_size": 63488 00:08:01.063 }, 00:08:01.063 { 00:08:01.063 "name": "pt2", 00:08:01.063 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:01.063 "is_configured": true, 00:08:01.063 "data_offset": 2048, 00:08:01.063 "data_size": 63488 00:08:01.063 } 00:08:01.063 ] 00:08:01.063 } 00:08:01.063 } 00:08:01.063 }' 00:08:01.063 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:01.063 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:01.063 pt2' 00:08:01.063 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.324 [2024-12-05 18:58:18.758411] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=5a793038-9d08-4e17-aabd-0542827c168b 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 5a793038-9d08-4e17-aabd-0542827c168b ']' 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.324 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.324 [2024-12-05 18:58:18.806091] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:01.324 [2024-12-05 18:58:18.806156] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:01.325 [2024-12-05 18:58:18.806257] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:01.325 [2024-12-05 18:58:18.806354] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:01.325 [2024-12-05 18:58:18.806443] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.325 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:01.586 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.586 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:01.586 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:01.586 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:08:01.586 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:01.586 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:08:01.586 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:01.586 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:08:01.586 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:01.586 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:01.586 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.586 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.586 [2024-12-05 18:58:18.921906] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:01.586 [2024-12-05 18:58:18.923814] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:01.586 [2024-12-05 18:58:18.923902] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:01.586 [2024-12-05 18:58:18.923953] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:01.586 [2024-12-05 18:58:18.923969] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:01.586 [2024-12-05 18:58:18.923978] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:08:01.587 request: 00:08:01.587 { 00:08:01.587 "name": "raid_bdev1", 00:08:01.587 "raid_level": "raid1", 00:08:01.587 "base_bdevs": [ 00:08:01.587 "malloc1", 00:08:01.587 "malloc2" 00:08:01.587 ], 00:08:01.587 "superblock": false, 00:08:01.587 "method": "bdev_raid_create", 00:08:01.587 "req_id": 1 00:08:01.587 } 00:08:01.587 Got JSON-RPC error response 00:08:01.587 response: 00:08:01.587 { 00:08:01.587 "code": -17, 00:08:01.587 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:01.587 } 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.587 [2024-12-05 18:58:18.985787] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:01.587 [2024-12-05 18:58:18.985886] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:01.587 [2024-12-05 18:58:18.985926] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:01.587 [2024-12-05 18:58:18.985953] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:01.587 [2024-12-05 18:58:18.988220] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:01.587 [2024-12-05 18:58:18.988293] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:01.587 [2024-12-05 18:58:18.988385] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:01.587 [2024-12-05 18:58:18.988471] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:01.587 pt1 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:01.587 18:58:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.587 18:58:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:01.587 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:01.587 "name": "raid_bdev1", 00:08:01.587 "uuid": "5a793038-9d08-4e17-aabd-0542827c168b", 00:08:01.587 "strip_size_kb": 0, 00:08:01.587 "state": "configuring", 00:08:01.587 "raid_level": "raid1", 00:08:01.587 "superblock": true, 00:08:01.587 "num_base_bdevs": 2, 00:08:01.587 "num_base_bdevs_discovered": 1, 00:08:01.587 "num_base_bdevs_operational": 2, 00:08:01.587 "base_bdevs_list": [ 00:08:01.587 { 00:08:01.587 "name": "pt1", 00:08:01.587 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:01.587 "is_configured": true, 00:08:01.587 "data_offset": 2048, 00:08:01.587 "data_size": 63488 00:08:01.587 }, 00:08:01.587 { 00:08:01.587 "name": null, 00:08:01.587 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:01.587 "is_configured": false, 00:08:01.587 "data_offset": 2048, 00:08:01.587 "data_size": 63488 00:08:01.587 } 00:08:01.587 ] 00:08:01.587 }' 00:08:01.587 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:01.587 18:58:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.157 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:08:02.157 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:02.157 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.158 [2024-12-05 18:58:19.457014] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:02.158 [2024-12-05 18:58:19.457069] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:02.158 [2024-12-05 18:58:19.457089] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:02.158 [2024-12-05 18:58:19.457098] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:02.158 [2024-12-05 18:58:19.457469] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:02.158 [2024-12-05 18:58:19.457485] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:02.158 [2024-12-05 18:58:19.457549] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:02.158 [2024-12-05 18:58:19.457567] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:02.158 [2024-12-05 18:58:19.457653] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:02.158 [2024-12-05 18:58:19.457661] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:02.158 [2024-12-05 18:58:19.457916] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:02.158 [2024-12-05 18:58:19.458055] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:02.158 [2024-12-05 18:58:19.458068] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:02.158 [2024-12-05 18:58:19.458172] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:02.158 pt2 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:02.158 "name": "raid_bdev1", 00:08:02.158 "uuid": "5a793038-9d08-4e17-aabd-0542827c168b", 00:08:02.158 "strip_size_kb": 0, 00:08:02.158 "state": "online", 00:08:02.158 "raid_level": "raid1", 00:08:02.158 "superblock": true, 00:08:02.158 "num_base_bdevs": 2, 00:08:02.158 "num_base_bdevs_discovered": 2, 00:08:02.158 "num_base_bdevs_operational": 2, 00:08:02.158 "base_bdevs_list": [ 00:08:02.158 { 00:08:02.158 "name": "pt1", 00:08:02.158 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:02.158 "is_configured": true, 00:08:02.158 "data_offset": 2048, 00:08:02.158 "data_size": 63488 00:08:02.158 }, 00:08:02.158 { 00:08:02.158 "name": "pt2", 00:08:02.158 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:02.158 "is_configured": true, 00:08:02.158 "data_offset": 2048, 00:08:02.158 "data_size": 63488 00:08:02.158 } 00:08:02.158 ] 00:08:02.158 }' 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:02.158 18:58:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.418 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:02.418 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:02.418 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:02.418 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:02.418 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:02.418 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:02.418 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:02.418 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:02.418 18:58:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.418 18:58:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.418 [2024-12-05 18:58:19.900478] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:02.418 18:58:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.418 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:02.418 "name": "raid_bdev1", 00:08:02.418 "aliases": [ 00:08:02.418 "5a793038-9d08-4e17-aabd-0542827c168b" 00:08:02.418 ], 00:08:02.418 "product_name": "Raid Volume", 00:08:02.418 "block_size": 512, 00:08:02.418 "num_blocks": 63488, 00:08:02.418 "uuid": "5a793038-9d08-4e17-aabd-0542827c168b", 00:08:02.418 "assigned_rate_limits": { 00:08:02.418 "rw_ios_per_sec": 0, 00:08:02.418 "rw_mbytes_per_sec": 0, 00:08:02.418 "r_mbytes_per_sec": 0, 00:08:02.418 "w_mbytes_per_sec": 0 00:08:02.418 }, 00:08:02.418 "claimed": false, 00:08:02.418 "zoned": false, 00:08:02.418 "supported_io_types": { 00:08:02.418 "read": true, 00:08:02.418 "write": true, 00:08:02.418 "unmap": false, 00:08:02.418 "flush": false, 00:08:02.418 "reset": true, 00:08:02.418 "nvme_admin": false, 00:08:02.418 "nvme_io": false, 00:08:02.418 "nvme_io_md": false, 00:08:02.418 "write_zeroes": true, 00:08:02.418 "zcopy": false, 00:08:02.418 "get_zone_info": false, 00:08:02.418 "zone_management": false, 00:08:02.418 "zone_append": false, 00:08:02.418 "compare": false, 00:08:02.418 "compare_and_write": false, 00:08:02.418 "abort": false, 00:08:02.418 "seek_hole": false, 00:08:02.418 "seek_data": false, 00:08:02.418 "copy": false, 00:08:02.418 "nvme_iov_md": false 00:08:02.418 }, 00:08:02.418 "memory_domains": [ 00:08:02.418 { 00:08:02.418 "dma_device_id": "system", 00:08:02.418 "dma_device_type": 1 00:08:02.418 }, 00:08:02.418 { 00:08:02.418 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:02.418 "dma_device_type": 2 00:08:02.418 }, 00:08:02.418 { 00:08:02.418 "dma_device_id": "system", 00:08:02.418 "dma_device_type": 1 00:08:02.418 }, 00:08:02.418 { 00:08:02.418 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:02.418 "dma_device_type": 2 00:08:02.418 } 00:08:02.418 ], 00:08:02.418 "driver_specific": { 00:08:02.418 "raid": { 00:08:02.418 "uuid": "5a793038-9d08-4e17-aabd-0542827c168b", 00:08:02.418 "strip_size_kb": 0, 00:08:02.418 "state": "online", 00:08:02.418 "raid_level": "raid1", 00:08:02.418 "superblock": true, 00:08:02.418 "num_base_bdevs": 2, 00:08:02.418 "num_base_bdevs_discovered": 2, 00:08:02.418 "num_base_bdevs_operational": 2, 00:08:02.418 "base_bdevs_list": [ 00:08:02.418 { 00:08:02.418 "name": "pt1", 00:08:02.418 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:02.418 "is_configured": true, 00:08:02.418 "data_offset": 2048, 00:08:02.418 "data_size": 63488 00:08:02.418 }, 00:08:02.418 { 00:08:02.418 "name": "pt2", 00:08:02.418 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:02.418 "is_configured": true, 00:08:02.418 "data_offset": 2048, 00:08:02.418 "data_size": 63488 00:08:02.418 } 00:08:02.418 ] 00:08:02.418 } 00:08:02.418 } 00:08:02.418 }' 00:08:02.418 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:02.418 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:02.418 pt2' 00:08:02.677 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:02.677 18:58:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:02.677 [2024-12-05 18:58:20.100146] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 5a793038-9d08-4e17-aabd-0542827c168b '!=' 5a793038-9d08-4e17-aabd-0542827c168b ']' 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.677 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.677 [2024-12-05 18:58:20.147834] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:02.678 "name": "raid_bdev1", 00:08:02.678 "uuid": "5a793038-9d08-4e17-aabd-0542827c168b", 00:08:02.678 "strip_size_kb": 0, 00:08:02.678 "state": "online", 00:08:02.678 "raid_level": "raid1", 00:08:02.678 "superblock": true, 00:08:02.678 "num_base_bdevs": 2, 00:08:02.678 "num_base_bdevs_discovered": 1, 00:08:02.678 "num_base_bdevs_operational": 1, 00:08:02.678 "base_bdevs_list": [ 00:08:02.678 { 00:08:02.678 "name": null, 00:08:02.678 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:02.678 "is_configured": false, 00:08:02.678 "data_offset": 0, 00:08:02.678 "data_size": 63488 00:08:02.678 }, 00:08:02.678 { 00:08:02.678 "name": "pt2", 00:08:02.678 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:02.678 "is_configured": true, 00:08:02.678 "data_offset": 2048, 00:08:02.678 "data_size": 63488 00:08:02.678 } 00:08:02.678 ] 00:08:02.678 }' 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:02.678 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.247 [2024-12-05 18:58:20.539414] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:03.247 [2024-12-05 18:58:20.539485] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:03.247 [2024-12-05 18:58:20.539575] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:03.247 [2024-12-05 18:58:20.539649] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:03.247 [2024-12-05 18:58:20.539771] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=1 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.247 [2024-12-05 18:58:20.595312] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:03.247 [2024-12-05 18:58:20.595359] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:03.247 [2024-12-05 18:58:20.595391] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:03.247 [2024-12-05 18:58:20.595399] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:03.247 [2024-12-05 18:58:20.597628] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:03.247 [2024-12-05 18:58:20.597673] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:03.247 [2024-12-05 18:58:20.597739] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:03.247 [2024-12-05 18:58:20.597767] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:03.247 [2024-12-05 18:58:20.597841] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:03.247 [2024-12-05 18:58:20.597849] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:03.247 [2024-12-05 18:58:20.598087] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:03.247 [2024-12-05 18:58:20.598204] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:03.247 [2024-12-05 18:58:20.598214] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:03.247 [2024-12-05 18:58:20.598317] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:03.247 pt2 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:03.247 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:03.248 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:03.248 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:03.248 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:03.248 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:03.248 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:03.248 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:03.248 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:03.248 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.248 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.248 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.248 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:03.248 "name": "raid_bdev1", 00:08:03.248 "uuid": "5a793038-9d08-4e17-aabd-0542827c168b", 00:08:03.248 "strip_size_kb": 0, 00:08:03.248 "state": "online", 00:08:03.248 "raid_level": "raid1", 00:08:03.248 "superblock": true, 00:08:03.248 "num_base_bdevs": 2, 00:08:03.248 "num_base_bdevs_discovered": 1, 00:08:03.248 "num_base_bdevs_operational": 1, 00:08:03.248 "base_bdevs_list": [ 00:08:03.248 { 00:08:03.248 "name": null, 00:08:03.248 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:03.248 "is_configured": false, 00:08:03.248 "data_offset": 2048, 00:08:03.248 "data_size": 63488 00:08:03.248 }, 00:08:03.248 { 00:08:03.248 "name": "pt2", 00:08:03.248 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:03.248 "is_configured": true, 00:08:03.248 "data_offset": 2048, 00:08:03.248 "data_size": 63488 00:08:03.248 } 00:08:03.248 ] 00:08:03.248 }' 00:08:03.248 18:58:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:03.248 18:58:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.508 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:03.508 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.508 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.508 [2024-12-05 18:58:21.030575] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:03.508 [2024-12-05 18:58:21.030646] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:03.508 [2024-12-05 18:58:21.030736] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:03.508 [2024-12-05 18:58:21.030795] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:03.508 [2024-12-05 18:58:21.030858] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:03.508 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.508 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:03.508 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:08:03.508 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.508 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.508 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.768 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:08:03.768 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:08:03.768 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:08:03.768 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:03.768 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.768 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.768 [2024-12-05 18:58:21.090465] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:03.768 [2024-12-05 18:58:21.090525] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:03.768 [2024-12-05 18:58:21.090543] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:08:03.768 [2024-12-05 18:58:21.090555] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:03.768 [2024-12-05 18:58:21.092705] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:03.768 [2024-12-05 18:58:21.092745] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:03.768 [2024-12-05 18:58:21.092807] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:03.768 [2024-12-05 18:58:21.092839] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:03.768 [2024-12-05 18:58:21.092932] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:08:03.768 [2024-12-05 18:58:21.092952] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:03.768 [2024-12-05 18:58:21.092972] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:08:03.768 [2024-12-05 18:58:21.093000] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:03.768 [2024-12-05 18:58:21.093064] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:08:03.768 [2024-12-05 18:58:21.093076] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:03.768 [2024-12-05 18:58:21.093302] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:08:03.768 [2024-12-05 18:58:21.093416] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:08:03.768 [2024-12-05 18:58:21.093425] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:08:03.768 [2024-12-05 18:58:21.093528] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:03.768 pt1 00:08:03.768 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.768 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:03.769 "name": "raid_bdev1", 00:08:03.769 "uuid": "5a793038-9d08-4e17-aabd-0542827c168b", 00:08:03.769 "strip_size_kb": 0, 00:08:03.769 "state": "online", 00:08:03.769 "raid_level": "raid1", 00:08:03.769 "superblock": true, 00:08:03.769 "num_base_bdevs": 2, 00:08:03.769 "num_base_bdevs_discovered": 1, 00:08:03.769 "num_base_bdevs_operational": 1, 00:08:03.769 "base_bdevs_list": [ 00:08:03.769 { 00:08:03.769 "name": null, 00:08:03.769 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:03.769 "is_configured": false, 00:08:03.769 "data_offset": 2048, 00:08:03.769 "data_size": 63488 00:08:03.769 }, 00:08:03.769 { 00:08:03.769 "name": "pt2", 00:08:03.769 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:03.769 "is_configured": true, 00:08:03.769 "data_offset": 2048, 00:08:03.769 "data_size": 63488 00:08:03.769 } 00:08:03.769 ] 00:08:03.769 }' 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:03.769 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.029 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:08:04.029 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.029 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.029 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:08:04.029 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.029 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:08:04.029 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:04.029 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:04.029 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.029 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:08:04.029 [2024-12-05 18:58:21.577855] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:04.029 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:04.289 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 5a793038-9d08-4e17-aabd-0542827c168b '!=' 5a793038-9d08-4e17-aabd-0542827c168b ']' 00:08:04.289 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 74148 00:08:04.289 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 74148 ']' 00:08:04.289 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 74148 00:08:04.289 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:04.289 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:04.289 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74148 00:08:04.289 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:04.289 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:04.289 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74148' 00:08:04.289 killing process with pid 74148 00:08:04.289 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 74148 00:08:04.289 [2024-12-05 18:58:21.664409] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:04.289 [2024-12-05 18:58:21.664532] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:04.289 [2024-12-05 18:58:21.664602] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to fr 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 74148 00:08:04.289 ee all in destruct 00:08:04.289 [2024-12-05 18:58:21.664646] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:08:04.289 [2024-12-05 18:58:21.686916] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:04.549 18:58:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:04.549 00:08:04.549 real 0m4.792s 00:08:04.549 user 0m7.905s 00:08:04.549 sys 0m0.908s 00:08:04.549 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:04.549 18:58:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.549 ************************************ 00:08:04.549 END TEST raid_superblock_test 00:08:04.549 ************************************ 00:08:04.549 18:58:21 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 2 read 00:08:04.549 18:58:21 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:04.549 18:58:21 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:04.549 18:58:21 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:04.549 ************************************ 00:08:04.549 START TEST raid_read_error_test 00:08:04.549 ************************************ 00:08:04.549 18:58:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 2 read 00:08:04.549 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:08:04.549 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:04.549 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:04.549 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.UU9HLSLEmz 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74462 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74462 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 74462 ']' 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:04.550 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:04.550 18:58:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.550 [2024-12-05 18:58:22.075623] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:08:04.550 [2024-12-05 18:58:22.075854] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74462 ] 00:08:04.810 [2024-12-05 18:58:22.220060] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:04.810 [2024-12-05 18:58:22.244772] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:04.810 [2024-12-05 18:58:22.286795] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:04.810 [2024-12-05 18:58:22.286830] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.380 BaseBdev1_malloc 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.380 true 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.380 [2024-12-05 18:58:22.925961] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:05.380 [2024-12-05 18:58:22.926074] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:05.380 [2024-12-05 18:58:22.926097] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:05.380 [2024-12-05 18:58:22.926106] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:05.380 [2024-12-05 18:58:22.928260] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:05.380 [2024-12-05 18:58:22.928301] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:05.380 BaseBdev1 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.380 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.640 BaseBdev2_malloc 00:08:05.640 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.640 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:05.640 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.640 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.640 true 00:08:05.640 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.640 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:05.640 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.640 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.640 [2024-12-05 18:58:22.966329] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:05.640 [2024-12-05 18:58:22.966376] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:05.640 [2024-12-05 18:58:22.966409] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:05.640 [2024-12-05 18:58:22.966424] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:05.640 [2024-12-05 18:58:22.968500] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:05.640 [2024-12-05 18:58:22.968548] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:05.640 BaseBdev2 00:08:05.640 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.640 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:05.640 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.640 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.640 [2024-12-05 18:58:22.978350] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:05.640 [2024-12-05 18:58:22.980233] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:05.640 [2024-12-05 18:58:22.980463] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:05.641 [2024-12-05 18:58:22.980498] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:05.641 [2024-12-05 18:58:22.980792] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:05.641 [2024-12-05 18:58:22.980977] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:05.641 [2024-12-05 18:58:22.981024] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:05.641 [2024-12-05 18:58:22.981197] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:05.641 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.641 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:05.641 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:05.641 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:05.641 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:05.641 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:05.641 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:05.641 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:05.641 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:05.641 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:05.641 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:05.641 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:05.641 18:58:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:05.641 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:05.641 18:58:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.641 18:58:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:05.641 18:58:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:05.641 "name": "raid_bdev1", 00:08:05.641 "uuid": "f0692395-ad84-4e8d-81df-1b9a729956c2", 00:08:05.641 "strip_size_kb": 0, 00:08:05.641 "state": "online", 00:08:05.641 "raid_level": "raid1", 00:08:05.641 "superblock": true, 00:08:05.641 "num_base_bdevs": 2, 00:08:05.641 "num_base_bdevs_discovered": 2, 00:08:05.641 "num_base_bdevs_operational": 2, 00:08:05.641 "base_bdevs_list": [ 00:08:05.641 { 00:08:05.641 "name": "BaseBdev1", 00:08:05.641 "uuid": "867f682d-7393-5ffd-98bc-ae40ce5812db", 00:08:05.641 "is_configured": true, 00:08:05.641 "data_offset": 2048, 00:08:05.641 "data_size": 63488 00:08:05.641 }, 00:08:05.641 { 00:08:05.641 "name": "BaseBdev2", 00:08:05.641 "uuid": "79650135-c9dc-5f9e-acd7-90d7ff419561", 00:08:05.641 "is_configured": true, 00:08:05.641 "data_offset": 2048, 00:08:05.641 "data_size": 63488 00:08:05.641 } 00:08:05.641 ] 00:08:05.641 }' 00:08:05.641 18:58:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:05.641 18:58:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.900 18:58:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:05.900 18:58:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:06.161 [2024-12-05 18:58:23.541802] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:08:07.102 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:07.103 "name": "raid_bdev1", 00:08:07.103 "uuid": "f0692395-ad84-4e8d-81df-1b9a729956c2", 00:08:07.103 "strip_size_kb": 0, 00:08:07.103 "state": "online", 00:08:07.103 "raid_level": "raid1", 00:08:07.103 "superblock": true, 00:08:07.103 "num_base_bdevs": 2, 00:08:07.103 "num_base_bdevs_discovered": 2, 00:08:07.103 "num_base_bdevs_operational": 2, 00:08:07.103 "base_bdevs_list": [ 00:08:07.103 { 00:08:07.103 "name": "BaseBdev1", 00:08:07.103 "uuid": "867f682d-7393-5ffd-98bc-ae40ce5812db", 00:08:07.103 "is_configured": true, 00:08:07.103 "data_offset": 2048, 00:08:07.103 "data_size": 63488 00:08:07.103 }, 00:08:07.103 { 00:08:07.103 "name": "BaseBdev2", 00:08:07.103 "uuid": "79650135-c9dc-5f9e-acd7-90d7ff419561", 00:08:07.103 "is_configured": true, 00:08:07.103 "data_offset": 2048, 00:08:07.103 "data_size": 63488 00:08:07.103 } 00:08:07.103 ] 00:08:07.103 }' 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:07.103 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.363 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:07.363 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:07.363 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.363 [2024-12-05 18:58:24.873044] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:07.363 [2024-12-05 18:58:24.873143] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:07.363 [2024-12-05 18:58:24.875732] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:07.364 [2024-12-05 18:58:24.875812] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:07.364 [2024-12-05 18:58:24.875916] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:07.364 [2024-12-05 18:58:24.876014] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:07.364 { 00:08:07.364 "results": [ 00:08:07.364 { 00:08:07.364 "job": "raid_bdev1", 00:08:07.364 "core_mask": "0x1", 00:08:07.364 "workload": "randrw", 00:08:07.364 "percentage": 50, 00:08:07.364 "status": "finished", 00:08:07.364 "queue_depth": 1, 00:08:07.364 "io_size": 131072, 00:08:07.364 "runtime": 1.332148, 00:08:07.364 "iops": 19962.496659530323, 00:08:07.364 "mibps": 2495.3120824412904, 00:08:07.364 "io_failed": 0, 00:08:07.364 "io_timeout": 0, 00:08:07.364 "avg_latency_us": 47.558383703102095, 00:08:07.364 "min_latency_us": 22.46986899563319, 00:08:07.364 "max_latency_us": 1366.5257641921398 00:08:07.364 } 00:08:07.364 ], 00:08:07.364 "core_count": 1 00:08:07.364 } 00:08:07.364 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:07.364 18:58:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74462 00:08:07.364 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 74462 ']' 00:08:07.364 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 74462 00:08:07.364 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:08:07.364 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:07.364 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74462 00:08:07.624 killing process with pid 74462 00:08:07.624 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:07.624 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:07.624 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74462' 00:08:07.624 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 74462 00:08:07.624 [2024-12-05 18:58:24.926102] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:07.624 18:58:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 74462 00:08:07.624 [2024-12-05 18:58:24.941663] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:07.624 18:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.UU9HLSLEmz 00:08:07.624 18:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:07.624 18:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:07.624 18:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:08:07.624 18:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:08:07.624 ************************************ 00:08:07.624 END TEST raid_read_error_test 00:08:07.624 ************************************ 00:08:07.624 18:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:07.624 18:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:07.624 18:58:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:08:07.624 00:08:07.624 real 0m3.181s 00:08:07.624 user 0m4.041s 00:08:07.624 sys 0m0.507s 00:08:07.624 18:58:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:07.624 18:58:25 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.883 18:58:25 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 2 write 00:08:07.883 18:58:25 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:07.883 18:58:25 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:07.883 18:58:25 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:07.883 ************************************ 00:08:07.883 START TEST raid_write_error_test 00:08:07.883 ************************************ 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 2 write 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.oRZtdO2qmD 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74591 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74591 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 74591 ']' 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:07.883 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:07.883 18:58:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.884 [2024-12-05 18:58:25.327307] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:08:07.884 [2024-12-05 18:58:25.327418] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74591 ] 00:08:08.143 [2024-12-05 18:58:25.480815] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:08.143 [2024-12-05 18:58:25.505832] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:08.143 [2024-12-05 18:58:25.549833] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:08.143 [2024-12-05 18:58:25.549885] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:08.711 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:08.711 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:08.711 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:08.711 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.712 BaseBdev1_malloc 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.712 true 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.712 [2024-12-05 18:58:26.185950] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:08.712 [2024-12-05 18:58:26.186064] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:08.712 [2024-12-05 18:58:26.186087] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:08.712 [2024-12-05 18:58:26.186096] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:08.712 [2024-12-05 18:58:26.188252] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:08.712 [2024-12-05 18:58:26.188302] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:08.712 BaseBdev1 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.712 BaseBdev2_malloc 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.712 true 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.712 [2024-12-05 18:58:26.226478] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:08.712 [2024-12-05 18:58:26.226526] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:08.712 [2024-12-05 18:58:26.226558] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:08.712 [2024-12-05 18:58:26.226574] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:08.712 [2024-12-05 18:58:26.228719] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:08.712 [2024-12-05 18:58:26.228809] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:08.712 BaseBdev2 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.712 [2024-12-05 18:58:26.238495] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:08.712 [2024-12-05 18:58:26.240303] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:08.712 [2024-12-05 18:58:26.240481] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:08.712 [2024-12-05 18:58:26.240495] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:08.712 [2024-12-05 18:58:26.240775] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:08.712 [2024-12-05 18:58:26.240930] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:08.712 [2024-12-05 18:58:26.240948] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:08.712 [2024-12-05 18:58:26.241058] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.712 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:08.972 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:08.972 "name": "raid_bdev1", 00:08:08.972 "uuid": "a9265134-9fe1-4d8f-89c8-f8d45edb00fe", 00:08:08.972 "strip_size_kb": 0, 00:08:08.972 "state": "online", 00:08:08.972 "raid_level": "raid1", 00:08:08.972 "superblock": true, 00:08:08.972 "num_base_bdevs": 2, 00:08:08.972 "num_base_bdevs_discovered": 2, 00:08:08.972 "num_base_bdevs_operational": 2, 00:08:08.972 "base_bdevs_list": [ 00:08:08.972 { 00:08:08.972 "name": "BaseBdev1", 00:08:08.972 "uuid": "66e5186d-d31b-5406-bb18-0374de24c6ee", 00:08:08.972 "is_configured": true, 00:08:08.972 "data_offset": 2048, 00:08:08.972 "data_size": 63488 00:08:08.972 }, 00:08:08.972 { 00:08:08.972 "name": "BaseBdev2", 00:08:08.972 "uuid": "a94a30e1-7074-5a73-b0df-7aefaef8ae03", 00:08:08.972 "is_configured": true, 00:08:08.972 "data_offset": 2048, 00:08:08.972 "data_size": 63488 00:08:08.972 } 00:08:08.972 ] 00:08:08.972 }' 00:08:08.972 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:08.972 18:58:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.232 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:09.233 18:58:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:09.233 [2024-12-05 18:58:26.742019] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.172 [2024-12-05 18:58:27.662611] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:08:10.172 [2024-12-05 18:58:27.662680] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:10.172 [2024-12-05 18:58:27.662923] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002a10 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=1 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.172 18:58:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.173 18:58:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.173 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:10.173 "name": "raid_bdev1", 00:08:10.173 "uuid": "a9265134-9fe1-4d8f-89c8-f8d45edb00fe", 00:08:10.173 "strip_size_kb": 0, 00:08:10.173 "state": "online", 00:08:10.173 "raid_level": "raid1", 00:08:10.173 "superblock": true, 00:08:10.173 "num_base_bdevs": 2, 00:08:10.173 "num_base_bdevs_discovered": 1, 00:08:10.173 "num_base_bdevs_operational": 1, 00:08:10.173 "base_bdevs_list": [ 00:08:10.173 { 00:08:10.173 "name": null, 00:08:10.173 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:10.173 "is_configured": false, 00:08:10.173 "data_offset": 0, 00:08:10.173 "data_size": 63488 00:08:10.173 }, 00:08:10.173 { 00:08:10.173 "name": "BaseBdev2", 00:08:10.173 "uuid": "a94a30e1-7074-5a73-b0df-7aefaef8ae03", 00:08:10.173 "is_configured": true, 00:08:10.173 "data_offset": 2048, 00:08:10.173 "data_size": 63488 00:08:10.173 } 00:08:10.173 ] 00:08:10.173 }' 00:08:10.173 18:58:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:10.173 18:58:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.742 18:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:10.742 18:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:10.742 18:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.742 [2024-12-05 18:58:28.112536] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:10.742 [2024-12-05 18:58:28.112651] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:10.742 [2024-12-05 18:58:28.115148] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:10.742 [2024-12-05 18:58:28.115255] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:10.742 [2024-12-05 18:58:28.115330] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:10.742 [2024-12-05 18:58:28.115383] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:10.742 { 00:08:10.742 "results": [ 00:08:10.742 { 00:08:10.742 "job": "raid_bdev1", 00:08:10.742 "core_mask": "0x1", 00:08:10.742 "workload": "randrw", 00:08:10.742 "percentage": 50, 00:08:10.742 "status": "finished", 00:08:10.742 "queue_depth": 1, 00:08:10.742 "io_size": 131072, 00:08:10.742 "runtime": 1.371521, 00:08:10.742 "iops": 23267.59852747424, 00:08:10.742 "mibps": 2908.44981593428, 00:08:10.742 "io_failed": 0, 00:08:10.742 "io_timeout": 0, 00:08:10.742 "avg_latency_us": 40.42508689288557, 00:08:10.742 "min_latency_us": 22.358078602620086, 00:08:10.742 "max_latency_us": 1359.3711790393013 00:08:10.742 } 00:08:10.742 ], 00:08:10.742 "core_count": 1 00:08:10.742 } 00:08:10.742 18:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:10.742 18:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74591 00:08:10.742 18:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 74591 ']' 00:08:10.742 18:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 74591 00:08:10.742 18:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:08:10.742 18:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:10.742 18:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74591 00:08:10.742 18:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:10.742 18:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:10.742 killing process with pid 74591 00:08:10.742 18:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74591' 00:08:10.742 18:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 74591 00:08:10.742 [2024-12-05 18:58:28.147705] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:10.742 18:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 74591 00:08:10.742 [2024-12-05 18:58:28.162438] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:11.006 18:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.oRZtdO2qmD 00:08:11.006 18:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:11.006 18:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:11.006 18:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:08:11.006 18:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:08:11.006 18:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:11.006 18:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:11.006 18:58:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:08:11.006 00:08:11.006 real 0m3.139s 00:08:11.006 user 0m3.997s 00:08:11.006 sys 0m0.473s 00:08:11.006 ************************************ 00:08:11.006 END TEST raid_write_error_test 00:08:11.006 ************************************ 00:08:11.006 18:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:11.006 18:58:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.006 18:58:28 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:08:11.006 18:58:28 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:11.006 18:58:28 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 3 false 00:08:11.006 18:58:28 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:11.006 18:58:28 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:11.006 18:58:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:11.006 ************************************ 00:08:11.006 START TEST raid_state_function_test 00:08:11.006 ************************************ 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 3 false 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:11.006 Process raid pid: 74718 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=74718 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 74718' 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 74718 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 74718 ']' 00:08:11.006 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:11.006 18:58:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.006 [2024-12-05 18:58:28.534097] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:08:11.006 [2024-12-05 18:58:28.534236] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:11.275 [2024-12-05 18:58:28.688111] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:11.275 [2024-12-05 18:58:28.713282] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:11.275 [2024-12-05 18:58:28.756227] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:11.275 [2024-12-05 18:58:28.756262] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:11.845 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:11.845 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:08:11.845 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:11.845 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:11.845 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.845 [2024-12-05 18:58:29.343701] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:11.845 [2024-12-05 18:58:29.343765] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:11.845 [2024-12-05 18:58:29.343775] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:11.846 [2024-12-05 18:58:29.343785] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:11.846 [2024-12-05 18:58:29.343790] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:11.846 [2024-12-05 18:58:29.343802] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:11.846 "name": "Existed_Raid", 00:08:11.846 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.846 "strip_size_kb": 64, 00:08:11.846 "state": "configuring", 00:08:11.846 "raid_level": "raid0", 00:08:11.846 "superblock": false, 00:08:11.846 "num_base_bdevs": 3, 00:08:11.846 "num_base_bdevs_discovered": 0, 00:08:11.846 "num_base_bdevs_operational": 3, 00:08:11.846 "base_bdevs_list": [ 00:08:11.846 { 00:08:11.846 "name": "BaseBdev1", 00:08:11.846 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.846 "is_configured": false, 00:08:11.846 "data_offset": 0, 00:08:11.846 "data_size": 0 00:08:11.846 }, 00:08:11.846 { 00:08:11.846 "name": "BaseBdev2", 00:08:11.846 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.846 "is_configured": false, 00:08:11.846 "data_offset": 0, 00:08:11.846 "data_size": 0 00:08:11.846 }, 00:08:11.846 { 00:08:11.846 "name": "BaseBdev3", 00:08:11.846 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.846 "is_configured": false, 00:08:11.846 "data_offset": 0, 00:08:11.846 "data_size": 0 00:08:11.846 } 00:08:11.846 ] 00:08:11.846 }' 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:11.846 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.416 [2024-12-05 18:58:29.766855] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:12.416 [2024-12-05 18:58:29.766940] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.416 [2024-12-05 18:58:29.778854] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:12.416 [2024-12-05 18:58:29.778949] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:12.416 [2024-12-05 18:58:29.778976] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:12.416 [2024-12-05 18:58:29.778998] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:12.416 [2024-12-05 18:58:29.779016] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:12.416 [2024-12-05 18:58:29.779036] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.416 [2024-12-05 18:58:29.799781] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:12.416 BaseBdev1 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.416 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.416 [ 00:08:12.416 { 00:08:12.416 "name": "BaseBdev1", 00:08:12.416 "aliases": [ 00:08:12.416 "0ae33dd7-164a-42c1-848a-b88c0959530b" 00:08:12.416 ], 00:08:12.416 "product_name": "Malloc disk", 00:08:12.416 "block_size": 512, 00:08:12.416 "num_blocks": 65536, 00:08:12.416 "uuid": "0ae33dd7-164a-42c1-848a-b88c0959530b", 00:08:12.416 "assigned_rate_limits": { 00:08:12.416 "rw_ios_per_sec": 0, 00:08:12.416 "rw_mbytes_per_sec": 0, 00:08:12.416 "r_mbytes_per_sec": 0, 00:08:12.416 "w_mbytes_per_sec": 0 00:08:12.416 }, 00:08:12.416 "claimed": true, 00:08:12.416 "claim_type": "exclusive_write", 00:08:12.416 "zoned": false, 00:08:12.416 "supported_io_types": { 00:08:12.416 "read": true, 00:08:12.416 "write": true, 00:08:12.416 "unmap": true, 00:08:12.416 "flush": true, 00:08:12.416 "reset": true, 00:08:12.416 "nvme_admin": false, 00:08:12.416 "nvme_io": false, 00:08:12.416 "nvme_io_md": false, 00:08:12.416 "write_zeroes": true, 00:08:12.416 "zcopy": true, 00:08:12.416 "get_zone_info": false, 00:08:12.416 "zone_management": false, 00:08:12.416 "zone_append": false, 00:08:12.416 "compare": false, 00:08:12.416 "compare_and_write": false, 00:08:12.416 "abort": true, 00:08:12.416 "seek_hole": false, 00:08:12.416 "seek_data": false, 00:08:12.416 "copy": true, 00:08:12.416 "nvme_iov_md": false 00:08:12.416 }, 00:08:12.416 "memory_domains": [ 00:08:12.416 { 00:08:12.416 "dma_device_id": "system", 00:08:12.416 "dma_device_type": 1 00:08:12.416 }, 00:08:12.416 { 00:08:12.416 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:12.416 "dma_device_type": 2 00:08:12.416 } 00:08:12.416 ], 00:08:12.416 "driver_specific": {} 00:08:12.416 } 00:08:12.416 ] 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:12.417 "name": "Existed_Raid", 00:08:12.417 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:12.417 "strip_size_kb": 64, 00:08:12.417 "state": "configuring", 00:08:12.417 "raid_level": "raid0", 00:08:12.417 "superblock": false, 00:08:12.417 "num_base_bdevs": 3, 00:08:12.417 "num_base_bdevs_discovered": 1, 00:08:12.417 "num_base_bdevs_operational": 3, 00:08:12.417 "base_bdevs_list": [ 00:08:12.417 { 00:08:12.417 "name": "BaseBdev1", 00:08:12.417 "uuid": "0ae33dd7-164a-42c1-848a-b88c0959530b", 00:08:12.417 "is_configured": true, 00:08:12.417 "data_offset": 0, 00:08:12.417 "data_size": 65536 00:08:12.417 }, 00:08:12.417 { 00:08:12.417 "name": "BaseBdev2", 00:08:12.417 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:12.417 "is_configured": false, 00:08:12.417 "data_offset": 0, 00:08:12.417 "data_size": 0 00:08:12.417 }, 00:08:12.417 { 00:08:12.417 "name": "BaseBdev3", 00:08:12.417 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:12.417 "is_configured": false, 00:08:12.417 "data_offset": 0, 00:08:12.417 "data_size": 0 00:08:12.417 } 00:08:12.417 ] 00:08:12.417 }' 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:12.417 18:58:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.677 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:12.677 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.677 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.677 [2024-12-05 18:58:30.203152] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:12.678 [2024-12-05 18:58:30.203243] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.678 [2024-12-05 18:58:30.215168] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:12.678 [2024-12-05 18:58:30.217006] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:12.678 [2024-12-05 18:58:30.217091] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:12.678 [2024-12-05 18:58:30.217106] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:12.678 [2024-12-05 18:58:30.217116] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:12.678 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.938 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:12.938 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:12.938 "name": "Existed_Raid", 00:08:12.938 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:12.938 "strip_size_kb": 64, 00:08:12.938 "state": "configuring", 00:08:12.938 "raid_level": "raid0", 00:08:12.938 "superblock": false, 00:08:12.938 "num_base_bdevs": 3, 00:08:12.938 "num_base_bdevs_discovered": 1, 00:08:12.938 "num_base_bdevs_operational": 3, 00:08:12.938 "base_bdevs_list": [ 00:08:12.938 { 00:08:12.938 "name": "BaseBdev1", 00:08:12.938 "uuid": "0ae33dd7-164a-42c1-848a-b88c0959530b", 00:08:12.938 "is_configured": true, 00:08:12.938 "data_offset": 0, 00:08:12.938 "data_size": 65536 00:08:12.938 }, 00:08:12.938 { 00:08:12.938 "name": "BaseBdev2", 00:08:12.938 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:12.938 "is_configured": false, 00:08:12.938 "data_offset": 0, 00:08:12.938 "data_size": 0 00:08:12.938 }, 00:08:12.938 { 00:08:12.938 "name": "BaseBdev3", 00:08:12.938 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:12.938 "is_configured": false, 00:08:12.938 "data_offset": 0, 00:08:12.938 "data_size": 0 00:08:12.938 } 00:08:12.938 ] 00:08:12.938 }' 00:08:12.938 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:12.938 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.199 [2024-12-05 18:58:30.673499] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:13.199 BaseBdev2 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.199 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.199 [ 00:08:13.199 { 00:08:13.199 "name": "BaseBdev2", 00:08:13.199 "aliases": [ 00:08:13.199 "465d61cf-d411-4888-9e22-37a8b9cbdabb" 00:08:13.199 ], 00:08:13.199 "product_name": "Malloc disk", 00:08:13.199 "block_size": 512, 00:08:13.199 "num_blocks": 65536, 00:08:13.199 "uuid": "465d61cf-d411-4888-9e22-37a8b9cbdabb", 00:08:13.199 "assigned_rate_limits": { 00:08:13.199 "rw_ios_per_sec": 0, 00:08:13.199 "rw_mbytes_per_sec": 0, 00:08:13.199 "r_mbytes_per_sec": 0, 00:08:13.199 "w_mbytes_per_sec": 0 00:08:13.199 }, 00:08:13.199 "claimed": true, 00:08:13.199 "claim_type": "exclusive_write", 00:08:13.199 "zoned": false, 00:08:13.200 "supported_io_types": { 00:08:13.200 "read": true, 00:08:13.200 "write": true, 00:08:13.200 "unmap": true, 00:08:13.200 "flush": true, 00:08:13.200 "reset": true, 00:08:13.200 "nvme_admin": false, 00:08:13.200 "nvme_io": false, 00:08:13.200 "nvme_io_md": false, 00:08:13.200 "write_zeroes": true, 00:08:13.200 "zcopy": true, 00:08:13.200 "get_zone_info": false, 00:08:13.200 "zone_management": false, 00:08:13.200 "zone_append": false, 00:08:13.200 "compare": false, 00:08:13.200 "compare_and_write": false, 00:08:13.200 "abort": true, 00:08:13.200 "seek_hole": false, 00:08:13.200 "seek_data": false, 00:08:13.200 "copy": true, 00:08:13.200 "nvme_iov_md": false 00:08:13.200 }, 00:08:13.200 "memory_domains": [ 00:08:13.200 { 00:08:13.200 "dma_device_id": "system", 00:08:13.200 "dma_device_type": 1 00:08:13.200 }, 00:08:13.200 { 00:08:13.200 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:13.200 "dma_device_type": 2 00:08:13.200 } 00:08:13.200 ], 00:08:13.200 "driver_specific": {} 00:08:13.200 } 00:08:13.200 ] 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:13.200 "name": "Existed_Raid", 00:08:13.200 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:13.200 "strip_size_kb": 64, 00:08:13.200 "state": "configuring", 00:08:13.200 "raid_level": "raid0", 00:08:13.200 "superblock": false, 00:08:13.200 "num_base_bdevs": 3, 00:08:13.200 "num_base_bdevs_discovered": 2, 00:08:13.200 "num_base_bdevs_operational": 3, 00:08:13.200 "base_bdevs_list": [ 00:08:13.200 { 00:08:13.200 "name": "BaseBdev1", 00:08:13.200 "uuid": "0ae33dd7-164a-42c1-848a-b88c0959530b", 00:08:13.200 "is_configured": true, 00:08:13.200 "data_offset": 0, 00:08:13.200 "data_size": 65536 00:08:13.200 }, 00:08:13.200 { 00:08:13.200 "name": "BaseBdev2", 00:08:13.200 "uuid": "465d61cf-d411-4888-9e22-37a8b9cbdabb", 00:08:13.200 "is_configured": true, 00:08:13.200 "data_offset": 0, 00:08:13.200 "data_size": 65536 00:08:13.200 }, 00:08:13.200 { 00:08:13.200 "name": "BaseBdev3", 00:08:13.200 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:13.200 "is_configured": false, 00:08:13.200 "data_offset": 0, 00:08:13.200 "data_size": 0 00:08:13.200 } 00:08:13.200 ] 00:08:13.200 }' 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:13.200 18:58:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.771 [2024-12-05 18:58:31.131769] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:13.771 [2024-12-05 18:58:31.131907] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:13.771 [2024-12-05 18:58:31.131940] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:13.771 [2024-12-05 18:58:31.132274] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:13.771 [2024-12-05 18:58:31.132444] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:13.771 [2024-12-05 18:58:31.132457] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:13.771 [2024-12-05 18:58:31.132735] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:13.771 BaseBdev3 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.771 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.771 [ 00:08:13.771 { 00:08:13.771 "name": "BaseBdev3", 00:08:13.771 "aliases": [ 00:08:13.771 "87850148-1961-4591-9a73-a99e2be2f347" 00:08:13.771 ], 00:08:13.771 "product_name": "Malloc disk", 00:08:13.771 "block_size": 512, 00:08:13.771 "num_blocks": 65536, 00:08:13.771 "uuid": "87850148-1961-4591-9a73-a99e2be2f347", 00:08:13.771 "assigned_rate_limits": { 00:08:13.771 "rw_ios_per_sec": 0, 00:08:13.771 "rw_mbytes_per_sec": 0, 00:08:13.771 "r_mbytes_per_sec": 0, 00:08:13.771 "w_mbytes_per_sec": 0 00:08:13.771 }, 00:08:13.771 "claimed": true, 00:08:13.771 "claim_type": "exclusive_write", 00:08:13.771 "zoned": false, 00:08:13.771 "supported_io_types": { 00:08:13.771 "read": true, 00:08:13.771 "write": true, 00:08:13.771 "unmap": true, 00:08:13.771 "flush": true, 00:08:13.771 "reset": true, 00:08:13.771 "nvme_admin": false, 00:08:13.771 "nvme_io": false, 00:08:13.771 "nvme_io_md": false, 00:08:13.771 "write_zeroes": true, 00:08:13.771 "zcopy": true, 00:08:13.772 "get_zone_info": false, 00:08:13.772 "zone_management": false, 00:08:13.772 "zone_append": false, 00:08:13.772 "compare": false, 00:08:13.772 "compare_and_write": false, 00:08:13.772 "abort": true, 00:08:13.772 "seek_hole": false, 00:08:13.772 "seek_data": false, 00:08:13.772 "copy": true, 00:08:13.772 "nvme_iov_md": false 00:08:13.772 }, 00:08:13.772 "memory_domains": [ 00:08:13.772 { 00:08:13.772 "dma_device_id": "system", 00:08:13.772 "dma_device_type": 1 00:08:13.772 }, 00:08:13.772 { 00:08:13.772 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:13.772 "dma_device_type": 2 00:08:13.772 } 00:08:13.772 ], 00:08:13.772 "driver_specific": {} 00:08:13.772 } 00:08:13.772 ] 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:13.772 "name": "Existed_Raid", 00:08:13.772 "uuid": "8aa6cb6a-a8db-4a15-a418-777c24053a98", 00:08:13.772 "strip_size_kb": 64, 00:08:13.772 "state": "online", 00:08:13.772 "raid_level": "raid0", 00:08:13.772 "superblock": false, 00:08:13.772 "num_base_bdevs": 3, 00:08:13.772 "num_base_bdevs_discovered": 3, 00:08:13.772 "num_base_bdevs_operational": 3, 00:08:13.772 "base_bdevs_list": [ 00:08:13.772 { 00:08:13.772 "name": "BaseBdev1", 00:08:13.772 "uuid": "0ae33dd7-164a-42c1-848a-b88c0959530b", 00:08:13.772 "is_configured": true, 00:08:13.772 "data_offset": 0, 00:08:13.772 "data_size": 65536 00:08:13.772 }, 00:08:13.772 { 00:08:13.772 "name": "BaseBdev2", 00:08:13.772 "uuid": "465d61cf-d411-4888-9e22-37a8b9cbdabb", 00:08:13.772 "is_configured": true, 00:08:13.772 "data_offset": 0, 00:08:13.772 "data_size": 65536 00:08:13.772 }, 00:08:13.772 { 00:08:13.772 "name": "BaseBdev3", 00:08:13.772 "uuid": "87850148-1961-4591-9a73-a99e2be2f347", 00:08:13.772 "is_configured": true, 00:08:13.772 "data_offset": 0, 00:08:13.772 "data_size": 65536 00:08:13.772 } 00:08:13.772 ] 00:08:13.772 }' 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:13.772 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.031 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:14.031 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:14.031 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:14.031 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:14.031 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:14.031 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:14.031 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:14.031 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:14.031 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.031 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.031 [2024-12-05 18:58:31.575490] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:14.290 "name": "Existed_Raid", 00:08:14.290 "aliases": [ 00:08:14.290 "8aa6cb6a-a8db-4a15-a418-777c24053a98" 00:08:14.290 ], 00:08:14.290 "product_name": "Raid Volume", 00:08:14.290 "block_size": 512, 00:08:14.290 "num_blocks": 196608, 00:08:14.290 "uuid": "8aa6cb6a-a8db-4a15-a418-777c24053a98", 00:08:14.290 "assigned_rate_limits": { 00:08:14.290 "rw_ios_per_sec": 0, 00:08:14.290 "rw_mbytes_per_sec": 0, 00:08:14.290 "r_mbytes_per_sec": 0, 00:08:14.290 "w_mbytes_per_sec": 0 00:08:14.290 }, 00:08:14.290 "claimed": false, 00:08:14.290 "zoned": false, 00:08:14.290 "supported_io_types": { 00:08:14.290 "read": true, 00:08:14.290 "write": true, 00:08:14.290 "unmap": true, 00:08:14.290 "flush": true, 00:08:14.290 "reset": true, 00:08:14.290 "nvme_admin": false, 00:08:14.290 "nvme_io": false, 00:08:14.290 "nvme_io_md": false, 00:08:14.290 "write_zeroes": true, 00:08:14.290 "zcopy": false, 00:08:14.290 "get_zone_info": false, 00:08:14.290 "zone_management": false, 00:08:14.290 "zone_append": false, 00:08:14.290 "compare": false, 00:08:14.290 "compare_and_write": false, 00:08:14.290 "abort": false, 00:08:14.290 "seek_hole": false, 00:08:14.290 "seek_data": false, 00:08:14.290 "copy": false, 00:08:14.290 "nvme_iov_md": false 00:08:14.290 }, 00:08:14.290 "memory_domains": [ 00:08:14.290 { 00:08:14.290 "dma_device_id": "system", 00:08:14.290 "dma_device_type": 1 00:08:14.290 }, 00:08:14.290 { 00:08:14.290 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:14.290 "dma_device_type": 2 00:08:14.290 }, 00:08:14.290 { 00:08:14.290 "dma_device_id": "system", 00:08:14.290 "dma_device_type": 1 00:08:14.290 }, 00:08:14.290 { 00:08:14.290 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:14.290 "dma_device_type": 2 00:08:14.290 }, 00:08:14.290 { 00:08:14.290 "dma_device_id": "system", 00:08:14.290 "dma_device_type": 1 00:08:14.290 }, 00:08:14.290 { 00:08:14.290 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:14.290 "dma_device_type": 2 00:08:14.290 } 00:08:14.290 ], 00:08:14.290 "driver_specific": { 00:08:14.290 "raid": { 00:08:14.290 "uuid": "8aa6cb6a-a8db-4a15-a418-777c24053a98", 00:08:14.290 "strip_size_kb": 64, 00:08:14.290 "state": "online", 00:08:14.290 "raid_level": "raid0", 00:08:14.290 "superblock": false, 00:08:14.290 "num_base_bdevs": 3, 00:08:14.290 "num_base_bdevs_discovered": 3, 00:08:14.290 "num_base_bdevs_operational": 3, 00:08:14.290 "base_bdevs_list": [ 00:08:14.290 { 00:08:14.290 "name": "BaseBdev1", 00:08:14.290 "uuid": "0ae33dd7-164a-42c1-848a-b88c0959530b", 00:08:14.290 "is_configured": true, 00:08:14.290 "data_offset": 0, 00:08:14.290 "data_size": 65536 00:08:14.290 }, 00:08:14.290 { 00:08:14.290 "name": "BaseBdev2", 00:08:14.290 "uuid": "465d61cf-d411-4888-9e22-37a8b9cbdabb", 00:08:14.290 "is_configured": true, 00:08:14.290 "data_offset": 0, 00:08:14.290 "data_size": 65536 00:08:14.290 }, 00:08:14.290 { 00:08:14.290 "name": "BaseBdev3", 00:08:14.290 "uuid": "87850148-1961-4591-9a73-a99e2be2f347", 00:08:14.290 "is_configured": true, 00:08:14.290 "data_offset": 0, 00:08:14.290 "data_size": 65536 00:08:14.290 } 00:08:14.290 ] 00:08:14.290 } 00:08:14.290 } 00:08:14.290 }' 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:14.290 BaseBdev2 00:08:14.290 BaseBdev3' 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:14.290 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.548 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.549 [2024-12-05 18:58:31.874776] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:14.549 [2024-12-05 18:58:31.874804] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:14.549 [2024-12-05 18:58:31.874862] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:14.549 "name": "Existed_Raid", 00:08:14.549 "uuid": "8aa6cb6a-a8db-4a15-a418-777c24053a98", 00:08:14.549 "strip_size_kb": 64, 00:08:14.549 "state": "offline", 00:08:14.549 "raid_level": "raid0", 00:08:14.549 "superblock": false, 00:08:14.549 "num_base_bdevs": 3, 00:08:14.549 "num_base_bdevs_discovered": 2, 00:08:14.549 "num_base_bdevs_operational": 2, 00:08:14.549 "base_bdevs_list": [ 00:08:14.549 { 00:08:14.549 "name": null, 00:08:14.549 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:14.549 "is_configured": false, 00:08:14.549 "data_offset": 0, 00:08:14.549 "data_size": 65536 00:08:14.549 }, 00:08:14.549 { 00:08:14.549 "name": "BaseBdev2", 00:08:14.549 "uuid": "465d61cf-d411-4888-9e22-37a8b9cbdabb", 00:08:14.549 "is_configured": true, 00:08:14.549 "data_offset": 0, 00:08:14.549 "data_size": 65536 00:08:14.549 }, 00:08:14.549 { 00:08:14.549 "name": "BaseBdev3", 00:08:14.549 "uuid": "87850148-1961-4591-9a73-a99e2be2f347", 00:08:14.549 "is_configured": true, 00:08:14.549 "data_offset": 0, 00:08:14.549 "data_size": 65536 00:08:14.549 } 00:08:14.549 ] 00:08:14.549 }' 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:14.549 18:58:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.807 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:14.807 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:14.807 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:14.807 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:14.807 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.807 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:14.807 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.066 [2024-12-05 18:58:32.369247] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.066 [2024-12-05 18:58:32.436253] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:15.066 [2024-12-05 18:58:32.436348] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.066 BaseBdev2 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.066 [ 00:08:15.066 { 00:08:15.066 "name": "BaseBdev2", 00:08:15.066 "aliases": [ 00:08:15.066 "bc765f72-919c-49e9-b7a0-3e7906896ad9" 00:08:15.066 ], 00:08:15.066 "product_name": "Malloc disk", 00:08:15.066 "block_size": 512, 00:08:15.066 "num_blocks": 65536, 00:08:15.066 "uuid": "bc765f72-919c-49e9-b7a0-3e7906896ad9", 00:08:15.066 "assigned_rate_limits": { 00:08:15.066 "rw_ios_per_sec": 0, 00:08:15.066 "rw_mbytes_per_sec": 0, 00:08:15.066 "r_mbytes_per_sec": 0, 00:08:15.066 "w_mbytes_per_sec": 0 00:08:15.066 }, 00:08:15.066 "claimed": false, 00:08:15.066 "zoned": false, 00:08:15.066 "supported_io_types": { 00:08:15.066 "read": true, 00:08:15.066 "write": true, 00:08:15.066 "unmap": true, 00:08:15.066 "flush": true, 00:08:15.066 "reset": true, 00:08:15.066 "nvme_admin": false, 00:08:15.066 "nvme_io": false, 00:08:15.066 "nvme_io_md": false, 00:08:15.066 "write_zeroes": true, 00:08:15.066 "zcopy": true, 00:08:15.066 "get_zone_info": false, 00:08:15.066 "zone_management": false, 00:08:15.066 "zone_append": false, 00:08:15.066 "compare": false, 00:08:15.066 "compare_and_write": false, 00:08:15.066 "abort": true, 00:08:15.066 "seek_hole": false, 00:08:15.066 "seek_data": false, 00:08:15.066 "copy": true, 00:08:15.066 "nvme_iov_md": false 00:08:15.066 }, 00:08:15.066 "memory_domains": [ 00:08:15.066 { 00:08:15.066 "dma_device_id": "system", 00:08:15.066 "dma_device_type": 1 00:08:15.066 }, 00:08:15.066 { 00:08:15.066 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:15.066 "dma_device_type": 2 00:08:15.066 } 00:08:15.066 ], 00:08:15.066 "driver_specific": {} 00:08:15.066 } 00:08:15.066 ] 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.066 BaseBdev3 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.066 [ 00:08:15.066 { 00:08:15.066 "name": "BaseBdev3", 00:08:15.066 "aliases": [ 00:08:15.066 "f8139e46-8072-4d97-86d5-8719ca37ed5d" 00:08:15.066 ], 00:08:15.066 "product_name": "Malloc disk", 00:08:15.066 "block_size": 512, 00:08:15.066 "num_blocks": 65536, 00:08:15.066 "uuid": "f8139e46-8072-4d97-86d5-8719ca37ed5d", 00:08:15.066 "assigned_rate_limits": { 00:08:15.066 "rw_ios_per_sec": 0, 00:08:15.066 "rw_mbytes_per_sec": 0, 00:08:15.066 "r_mbytes_per_sec": 0, 00:08:15.066 "w_mbytes_per_sec": 0 00:08:15.066 }, 00:08:15.066 "claimed": false, 00:08:15.066 "zoned": false, 00:08:15.066 "supported_io_types": { 00:08:15.066 "read": true, 00:08:15.066 "write": true, 00:08:15.066 "unmap": true, 00:08:15.066 "flush": true, 00:08:15.066 "reset": true, 00:08:15.066 "nvme_admin": false, 00:08:15.066 "nvme_io": false, 00:08:15.066 "nvme_io_md": false, 00:08:15.066 "write_zeroes": true, 00:08:15.066 "zcopy": true, 00:08:15.066 "get_zone_info": false, 00:08:15.066 "zone_management": false, 00:08:15.066 "zone_append": false, 00:08:15.066 "compare": false, 00:08:15.066 "compare_and_write": false, 00:08:15.066 "abort": true, 00:08:15.066 "seek_hole": false, 00:08:15.066 "seek_data": false, 00:08:15.066 "copy": true, 00:08:15.066 "nvme_iov_md": false 00:08:15.066 }, 00:08:15.066 "memory_domains": [ 00:08:15.066 { 00:08:15.066 "dma_device_id": "system", 00:08:15.066 "dma_device_type": 1 00:08:15.066 }, 00:08:15.066 { 00:08:15.066 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:15.066 "dma_device_type": 2 00:08:15.066 } 00:08:15.066 ], 00:08:15.066 "driver_specific": {} 00:08:15.066 } 00:08:15.066 ] 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.066 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.067 [2024-12-05 18:58:32.591421] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:15.067 [2024-12-05 18:58:32.591523] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:15.067 [2024-12-05 18:58:32.591565] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:15.067 [2024-12-05 18:58:32.593387] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.067 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.326 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:15.326 "name": "Existed_Raid", 00:08:15.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:15.326 "strip_size_kb": 64, 00:08:15.326 "state": "configuring", 00:08:15.326 "raid_level": "raid0", 00:08:15.326 "superblock": false, 00:08:15.326 "num_base_bdevs": 3, 00:08:15.326 "num_base_bdevs_discovered": 2, 00:08:15.326 "num_base_bdevs_operational": 3, 00:08:15.326 "base_bdevs_list": [ 00:08:15.326 { 00:08:15.326 "name": "BaseBdev1", 00:08:15.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:15.326 "is_configured": false, 00:08:15.326 "data_offset": 0, 00:08:15.326 "data_size": 0 00:08:15.326 }, 00:08:15.326 { 00:08:15.326 "name": "BaseBdev2", 00:08:15.326 "uuid": "bc765f72-919c-49e9-b7a0-3e7906896ad9", 00:08:15.326 "is_configured": true, 00:08:15.326 "data_offset": 0, 00:08:15.326 "data_size": 65536 00:08:15.326 }, 00:08:15.326 { 00:08:15.326 "name": "BaseBdev3", 00:08:15.326 "uuid": "f8139e46-8072-4d97-86d5-8719ca37ed5d", 00:08:15.326 "is_configured": true, 00:08:15.326 "data_offset": 0, 00:08:15.326 "data_size": 65536 00:08:15.326 } 00:08:15.326 ] 00:08:15.326 }' 00:08:15.326 18:58:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:15.326 18:58:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.586 [2024-12-05 18:58:33.034678] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:15.586 "name": "Existed_Raid", 00:08:15.586 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:15.586 "strip_size_kb": 64, 00:08:15.586 "state": "configuring", 00:08:15.586 "raid_level": "raid0", 00:08:15.586 "superblock": false, 00:08:15.586 "num_base_bdevs": 3, 00:08:15.586 "num_base_bdevs_discovered": 1, 00:08:15.586 "num_base_bdevs_operational": 3, 00:08:15.586 "base_bdevs_list": [ 00:08:15.586 { 00:08:15.586 "name": "BaseBdev1", 00:08:15.586 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:15.586 "is_configured": false, 00:08:15.586 "data_offset": 0, 00:08:15.586 "data_size": 0 00:08:15.586 }, 00:08:15.586 { 00:08:15.586 "name": null, 00:08:15.586 "uuid": "bc765f72-919c-49e9-b7a0-3e7906896ad9", 00:08:15.586 "is_configured": false, 00:08:15.586 "data_offset": 0, 00:08:15.586 "data_size": 65536 00:08:15.586 }, 00:08:15.586 { 00:08:15.586 "name": "BaseBdev3", 00:08:15.586 "uuid": "f8139e46-8072-4d97-86d5-8719ca37ed5d", 00:08:15.586 "is_configured": true, 00:08:15.586 "data_offset": 0, 00:08:15.586 "data_size": 65536 00:08:15.586 } 00:08:15.586 ] 00:08:15.586 }' 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:15.586 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.156 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.157 [2024-12-05 18:58:33.504906] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:16.157 BaseBdev1 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.157 [ 00:08:16.157 { 00:08:16.157 "name": "BaseBdev1", 00:08:16.157 "aliases": [ 00:08:16.157 "64eebdff-d224-4f27-b8d1-304fe00d0ece" 00:08:16.157 ], 00:08:16.157 "product_name": "Malloc disk", 00:08:16.157 "block_size": 512, 00:08:16.157 "num_blocks": 65536, 00:08:16.157 "uuid": "64eebdff-d224-4f27-b8d1-304fe00d0ece", 00:08:16.157 "assigned_rate_limits": { 00:08:16.157 "rw_ios_per_sec": 0, 00:08:16.157 "rw_mbytes_per_sec": 0, 00:08:16.157 "r_mbytes_per_sec": 0, 00:08:16.157 "w_mbytes_per_sec": 0 00:08:16.157 }, 00:08:16.157 "claimed": true, 00:08:16.157 "claim_type": "exclusive_write", 00:08:16.157 "zoned": false, 00:08:16.157 "supported_io_types": { 00:08:16.157 "read": true, 00:08:16.157 "write": true, 00:08:16.157 "unmap": true, 00:08:16.157 "flush": true, 00:08:16.157 "reset": true, 00:08:16.157 "nvme_admin": false, 00:08:16.157 "nvme_io": false, 00:08:16.157 "nvme_io_md": false, 00:08:16.157 "write_zeroes": true, 00:08:16.157 "zcopy": true, 00:08:16.157 "get_zone_info": false, 00:08:16.157 "zone_management": false, 00:08:16.157 "zone_append": false, 00:08:16.157 "compare": false, 00:08:16.157 "compare_and_write": false, 00:08:16.157 "abort": true, 00:08:16.157 "seek_hole": false, 00:08:16.157 "seek_data": false, 00:08:16.157 "copy": true, 00:08:16.157 "nvme_iov_md": false 00:08:16.157 }, 00:08:16.157 "memory_domains": [ 00:08:16.157 { 00:08:16.157 "dma_device_id": "system", 00:08:16.157 "dma_device_type": 1 00:08:16.157 }, 00:08:16.157 { 00:08:16.157 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:16.157 "dma_device_type": 2 00:08:16.157 } 00:08:16.157 ], 00:08:16.157 "driver_specific": {} 00:08:16.157 } 00:08:16.157 ] 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:16.157 "name": "Existed_Raid", 00:08:16.157 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.157 "strip_size_kb": 64, 00:08:16.157 "state": "configuring", 00:08:16.157 "raid_level": "raid0", 00:08:16.157 "superblock": false, 00:08:16.157 "num_base_bdevs": 3, 00:08:16.157 "num_base_bdevs_discovered": 2, 00:08:16.157 "num_base_bdevs_operational": 3, 00:08:16.157 "base_bdevs_list": [ 00:08:16.157 { 00:08:16.157 "name": "BaseBdev1", 00:08:16.157 "uuid": "64eebdff-d224-4f27-b8d1-304fe00d0ece", 00:08:16.157 "is_configured": true, 00:08:16.157 "data_offset": 0, 00:08:16.157 "data_size": 65536 00:08:16.157 }, 00:08:16.157 { 00:08:16.157 "name": null, 00:08:16.157 "uuid": "bc765f72-919c-49e9-b7a0-3e7906896ad9", 00:08:16.157 "is_configured": false, 00:08:16.157 "data_offset": 0, 00:08:16.157 "data_size": 65536 00:08:16.157 }, 00:08:16.157 { 00:08:16.157 "name": "BaseBdev3", 00:08:16.157 "uuid": "f8139e46-8072-4d97-86d5-8719ca37ed5d", 00:08:16.157 "is_configured": true, 00:08:16.157 "data_offset": 0, 00:08:16.157 "data_size": 65536 00:08:16.157 } 00:08:16.157 ] 00:08:16.157 }' 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:16.157 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.418 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:16.418 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.418 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.418 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:16.418 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.676 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:16.676 18:58:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:16.676 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.676 18:58:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.676 [2024-12-05 18:58:34.000113] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:16.676 "name": "Existed_Raid", 00:08:16.676 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.676 "strip_size_kb": 64, 00:08:16.676 "state": "configuring", 00:08:16.676 "raid_level": "raid0", 00:08:16.676 "superblock": false, 00:08:16.676 "num_base_bdevs": 3, 00:08:16.676 "num_base_bdevs_discovered": 1, 00:08:16.676 "num_base_bdevs_operational": 3, 00:08:16.676 "base_bdevs_list": [ 00:08:16.676 { 00:08:16.676 "name": "BaseBdev1", 00:08:16.676 "uuid": "64eebdff-d224-4f27-b8d1-304fe00d0ece", 00:08:16.676 "is_configured": true, 00:08:16.676 "data_offset": 0, 00:08:16.676 "data_size": 65536 00:08:16.676 }, 00:08:16.676 { 00:08:16.676 "name": null, 00:08:16.676 "uuid": "bc765f72-919c-49e9-b7a0-3e7906896ad9", 00:08:16.676 "is_configured": false, 00:08:16.676 "data_offset": 0, 00:08:16.676 "data_size": 65536 00:08:16.676 }, 00:08:16.676 { 00:08:16.676 "name": null, 00:08:16.676 "uuid": "f8139e46-8072-4d97-86d5-8719ca37ed5d", 00:08:16.676 "is_configured": false, 00:08:16.676 "data_offset": 0, 00:08:16.676 "data_size": 65536 00:08:16.676 } 00:08:16.676 ] 00:08:16.676 }' 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:16.676 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.935 [2024-12-05 18:58:34.463544] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:16.935 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.193 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:17.193 "name": "Existed_Raid", 00:08:17.193 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:17.193 "strip_size_kb": 64, 00:08:17.193 "state": "configuring", 00:08:17.193 "raid_level": "raid0", 00:08:17.193 "superblock": false, 00:08:17.193 "num_base_bdevs": 3, 00:08:17.193 "num_base_bdevs_discovered": 2, 00:08:17.193 "num_base_bdevs_operational": 3, 00:08:17.193 "base_bdevs_list": [ 00:08:17.193 { 00:08:17.193 "name": "BaseBdev1", 00:08:17.193 "uuid": "64eebdff-d224-4f27-b8d1-304fe00d0ece", 00:08:17.193 "is_configured": true, 00:08:17.193 "data_offset": 0, 00:08:17.193 "data_size": 65536 00:08:17.193 }, 00:08:17.193 { 00:08:17.193 "name": null, 00:08:17.193 "uuid": "bc765f72-919c-49e9-b7a0-3e7906896ad9", 00:08:17.193 "is_configured": false, 00:08:17.193 "data_offset": 0, 00:08:17.193 "data_size": 65536 00:08:17.193 }, 00:08:17.193 { 00:08:17.193 "name": "BaseBdev3", 00:08:17.193 "uuid": "f8139e46-8072-4d97-86d5-8719ca37ed5d", 00:08:17.193 "is_configured": true, 00:08:17.193 "data_offset": 0, 00:08:17.193 "data_size": 65536 00:08:17.193 } 00:08:17.193 ] 00:08:17.193 }' 00:08:17.193 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:17.193 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.452 [2024-12-05 18:58:34.978695] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.452 18:58:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.712 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.712 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:17.712 "name": "Existed_Raid", 00:08:17.712 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:17.712 "strip_size_kb": 64, 00:08:17.712 "state": "configuring", 00:08:17.712 "raid_level": "raid0", 00:08:17.712 "superblock": false, 00:08:17.712 "num_base_bdevs": 3, 00:08:17.712 "num_base_bdevs_discovered": 1, 00:08:17.712 "num_base_bdevs_operational": 3, 00:08:17.712 "base_bdevs_list": [ 00:08:17.712 { 00:08:17.712 "name": null, 00:08:17.712 "uuid": "64eebdff-d224-4f27-b8d1-304fe00d0ece", 00:08:17.712 "is_configured": false, 00:08:17.712 "data_offset": 0, 00:08:17.712 "data_size": 65536 00:08:17.712 }, 00:08:17.712 { 00:08:17.712 "name": null, 00:08:17.712 "uuid": "bc765f72-919c-49e9-b7a0-3e7906896ad9", 00:08:17.712 "is_configured": false, 00:08:17.712 "data_offset": 0, 00:08:17.712 "data_size": 65536 00:08:17.712 }, 00:08:17.712 { 00:08:17.712 "name": "BaseBdev3", 00:08:17.712 "uuid": "f8139e46-8072-4d97-86d5-8719ca37ed5d", 00:08:17.712 "is_configured": true, 00:08:17.712 "data_offset": 0, 00:08:17.712 "data_size": 65536 00:08:17.712 } 00:08:17.712 ] 00:08:17.712 }' 00:08:17.712 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:17.712 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.972 [2024-12-05 18:58:35.500279] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:17.972 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.231 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.231 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:18.231 "name": "Existed_Raid", 00:08:18.231 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:18.231 "strip_size_kb": 64, 00:08:18.231 "state": "configuring", 00:08:18.231 "raid_level": "raid0", 00:08:18.231 "superblock": false, 00:08:18.231 "num_base_bdevs": 3, 00:08:18.231 "num_base_bdevs_discovered": 2, 00:08:18.231 "num_base_bdevs_operational": 3, 00:08:18.231 "base_bdevs_list": [ 00:08:18.231 { 00:08:18.231 "name": null, 00:08:18.231 "uuid": "64eebdff-d224-4f27-b8d1-304fe00d0ece", 00:08:18.231 "is_configured": false, 00:08:18.231 "data_offset": 0, 00:08:18.231 "data_size": 65536 00:08:18.231 }, 00:08:18.231 { 00:08:18.231 "name": "BaseBdev2", 00:08:18.231 "uuid": "bc765f72-919c-49e9-b7a0-3e7906896ad9", 00:08:18.231 "is_configured": true, 00:08:18.231 "data_offset": 0, 00:08:18.231 "data_size": 65536 00:08:18.231 }, 00:08:18.231 { 00:08:18.231 "name": "BaseBdev3", 00:08:18.231 "uuid": "f8139e46-8072-4d97-86d5-8719ca37ed5d", 00:08:18.231 "is_configured": true, 00:08:18.231 "data_offset": 0, 00:08:18.231 "data_size": 65536 00:08:18.231 } 00:08:18.231 ] 00:08:18.231 }' 00:08:18.231 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:18.231 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.490 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:18.490 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:18.490 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.490 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.490 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.490 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:18.490 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:18.490 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.490 18:58:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.490 18:58:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:18.490 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.490 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 64eebdff-d224-4f27-b8d1-304fe00d0ece 00:08:18.490 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.490 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.750 [2024-12-05 18:58:36.050571] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:18.750 [2024-12-05 18:58:36.050689] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:18.750 [2024-12-05 18:58:36.050708] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:18.750 [2024-12-05 18:58:36.051012] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:18.750 [2024-12-05 18:58:36.051139] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:18.750 [2024-12-05 18:58:36.051149] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:18.750 [2024-12-05 18:58:36.051337] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:18.750 NewBaseBdev 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.750 [ 00:08:18.750 { 00:08:18.750 "name": "NewBaseBdev", 00:08:18.750 "aliases": [ 00:08:18.750 "64eebdff-d224-4f27-b8d1-304fe00d0ece" 00:08:18.750 ], 00:08:18.750 "product_name": "Malloc disk", 00:08:18.750 "block_size": 512, 00:08:18.750 "num_blocks": 65536, 00:08:18.750 "uuid": "64eebdff-d224-4f27-b8d1-304fe00d0ece", 00:08:18.750 "assigned_rate_limits": { 00:08:18.750 "rw_ios_per_sec": 0, 00:08:18.750 "rw_mbytes_per_sec": 0, 00:08:18.750 "r_mbytes_per_sec": 0, 00:08:18.750 "w_mbytes_per_sec": 0 00:08:18.750 }, 00:08:18.750 "claimed": true, 00:08:18.750 "claim_type": "exclusive_write", 00:08:18.750 "zoned": false, 00:08:18.750 "supported_io_types": { 00:08:18.750 "read": true, 00:08:18.750 "write": true, 00:08:18.750 "unmap": true, 00:08:18.750 "flush": true, 00:08:18.750 "reset": true, 00:08:18.750 "nvme_admin": false, 00:08:18.750 "nvme_io": false, 00:08:18.750 "nvme_io_md": false, 00:08:18.750 "write_zeroes": true, 00:08:18.750 "zcopy": true, 00:08:18.750 "get_zone_info": false, 00:08:18.750 "zone_management": false, 00:08:18.750 "zone_append": false, 00:08:18.750 "compare": false, 00:08:18.750 "compare_and_write": false, 00:08:18.750 "abort": true, 00:08:18.750 "seek_hole": false, 00:08:18.750 "seek_data": false, 00:08:18.750 "copy": true, 00:08:18.750 "nvme_iov_md": false 00:08:18.750 }, 00:08:18.750 "memory_domains": [ 00:08:18.750 { 00:08:18.750 "dma_device_id": "system", 00:08:18.750 "dma_device_type": 1 00:08:18.750 }, 00:08:18.750 { 00:08:18.750 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:18.750 "dma_device_type": 2 00:08:18.750 } 00:08:18.750 ], 00:08:18.750 "driver_specific": {} 00:08:18.750 } 00:08:18.750 ] 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:18.750 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:18.751 "name": "Existed_Raid", 00:08:18.751 "uuid": "df45b567-be79-4f7f-ab5a-44570ab08458", 00:08:18.751 "strip_size_kb": 64, 00:08:18.751 "state": "online", 00:08:18.751 "raid_level": "raid0", 00:08:18.751 "superblock": false, 00:08:18.751 "num_base_bdevs": 3, 00:08:18.751 "num_base_bdevs_discovered": 3, 00:08:18.751 "num_base_bdevs_operational": 3, 00:08:18.751 "base_bdevs_list": [ 00:08:18.751 { 00:08:18.751 "name": "NewBaseBdev", 00:08:18.751 "uuid": "64eebdff-d224-4f27-b8d1-304fe00d0ece", 00:08:18.751 "is_configured": true, 00:08:18.751 "data_offset": 0, 00:08:18.751 "data_size": 65536 00:08:18.751 }, 00:08:18.751 { 00:08:18.751 "name": "BaseBdev2", 00:08:18.751 "uuid": "bc765f72-919c-49e9-b7a0-3e7906896ad9", 00:08:18.751 "is_configured": true, 00:08:18.751 "data_offset": 0, 00:08:18.751 "data_size": 65536 00:08:18.751 }, 00:08:18.751 { 00:08:18.751 "name": "BaseBdev3", 00:08:18.751 "uuid": "f8139e46-8072-4d97-86d5-8719ca37ed5d", 00:08:18.751 "is_configured": true, 00:08:18.751 "data_offset": 0, 00:08:18.751 "data_size": 65536 00:08:18.751 } 00:08:18.751 ] 00:08:18.751 }' 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:18.751 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.011 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:19.011 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:19.011 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:19.011 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:19.011 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:19.011 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:19.011 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:19.011 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.011 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.011 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:19.011 [2024-12-05 18:58:36.494108] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:19.011 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.011 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:19.011 "name": "Existed_Raid", 00:08:19.011 "aliases": [ 00:08:19.011 "df45b567-be79-4f7f-ab5a-44570ab08458" 00:08:19.011 ], 00:08:19.011 "product_name": "Raid Volume", 00:08:19.011 "block_size": 512, 00:08:19.011 "num_blocks": 196608, 00:08:19.011 "uuid": "df45b567-be79-4f7f-ab5a-44570ab08458", 00:08:19.011 "assigned_rate_limits": { 00:08:19.011 "rw_ios_per_sec": 0, 00:08:19.011 "rw_mbytes_per_sec": 0, 00:08:19.011 "r_mbytes_per_sec": 0, 00:08:19.011 "w_mbytes_per_sec": 0 00:08:19.011 }, 00:08:19.011 "claimed": false, 00:08:19.011 "zoned": false, 00:08:19.011 "supported_io_types": { 00:08:19.011 "read": true, 00:08:19.011 "write": true, 00:08:19.011 "unmap": true, 00:08:19.011 "flush": true, 00:08:19.011 "reset": true, 00:08:19.011 "nvme_admin": false, 00:08:19.011 "nvme_io": false, 00:08:19.011 "nvme_io_md": false, 00:08:19.011 "write_zeroes": true, 00:08:19.011 "zcopy": false, 00:08:19.011 "get_zone_info": false, 00:08:19.011 "zone_management": false, 00:08:19.011 "zone_append": false, 00:08:19.011 "compare": false, 00:08:19.011 "compare_and_write": false, 00:08:19.011 "abort": false, 00:08:19.011 "seek_hole": false, 00:08:19.011 "seek_data": false, 00:08:19.011 "copy": false, 00:08:19.011 "nvme_iov_md": false 00:08:19.011 }, 00:08:19.011 "memory_domains": [ 00:08:19.011 { 00:08:19.011 "dma_device_id": "system", 00:08:19.011 "dma_device_type": 1 00:08:19.011 }, 00:08:19.011 { 00:08:19.011 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:19.011 "dma_device_type": 2 00:08:19.011 }, 00:08:19.011 { 00:08:19.011 "dma_device_id": "system", 00:08:19.011 "dma_device_type": 1 00:08:19.011 }, 00:08:19.011 { 00:08:19.011 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:19.011 "dma_device_type": 2 00:08:19.011 }, 00:08:19.011 { 00:08:19.011 "dma_device_id": "system", 00:08:19.011 "dma_device_type": 1 00:08:19.011 }, 00:08:19.011 { 00:08:19.011 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:19.011 "dma_device_type": 2 00:08:19.011 } 00:08:19.011 ], 00:08:19.011 "driver_specific": { 00:08:19.011 "raid": { 00:08:19.011 "uuid": "df45b567-be79-4f7f-ab5a-44570ab08458", 00:08:19.011 "strip_size_kb": 64, 00:08:19.011 "state": "online", 00:08:19.011 "raid_level": "raid0", 00:08:19.011 "superblock": false, 00:08:19.011 "num_base_bdevs": 3, 00:08:19.011 "num_base_bdevs_discovered": 3, 00:08:19.011 "num_base_bdevs_operational": 3, 00:08:19.011 "base_bdevs_list": [ 00:08:19.011 { 00:08:19.011 "name": "NewBaseBdev", 00:08:19.011 "uuid": "64eebdff-d224-4f27-b8d1-304fe00d0ece", 00:08:19.011 "is_configured": true, 00:08:19.011 "data_offset": 0, 00:08:19.011 "data_size": 65536 00:08:19.011 }, 00:08:19.011 { 00:08:19.011 "name": "BaseBdev2", 00:08:19.011 "uuid": "bc765f72-919c-49e9-b7a0-3e7906896ad9", 00:08:19.011 "is_configured": true, 00:08:19.011 "data_offset": 0, 00:08:19.011 "data_size": 65536 00:08:19.011 }, 00:08:19.011 { 00:08:19.011 "name": "BaseBdev3", 00:08:19.011 "uuid": "f8139e46-8072-4d97-86d5-8719ca37ed5d", 00:08:19.011 "is_configured": true, 00:08:19.011 "data_offset": 0, 00:08:19.011 "data_size": 65536 00:08:19.011 } 00:08:19.011 ] 00:08:19.011 } 00:08:19.011 } 00:08:19.011 }' 00:08:19.011 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:19.271 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:19.271 BaseBdev2 00:08:19.271 BaseBdev3' 00:08:19.271 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.272 [2024-12-05 18:58:36.749410] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:19.272 [2024-12-05 18:58:36.749435] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:19.272 [2024-12-05 18:58:36.749495] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:19.272 [2024-12-05 18:58:36.749541] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:19.272 [2024-12-05 18:58:36.749560] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 74718 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 74718 ']' 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 74718 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 74718 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 74718' 00:08:19.272 killing process with pid 74718 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 74718 00:08:19.272 [2024-12-05 18:58:36.800523] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:19.272 18:58:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 74718 00:08:19.532 [2024-12-05 18:58:36.831394] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:19.532 18:58:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:19.532 00:08:19.532 real 0m8.606s 00:08:19.532 user 0m14.724s 00:08:19.532 sys 0m1.679s 00:08:19.532 18:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:19.532 18:58:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.532 ************************************ 00:08:19.532 END TEST raid_state_function_test 00:08:19.532 ************************************ 00:08:19.792 18:58:37 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 3 true 00:08:19.792 18:58:37 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:19.792 18:58:37 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:19.792 18:58:37 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:19.792 ************************************ 00:08:19.792 START TEST raid_state_function_test_sb 00:08:19.792 ************************************ 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 3 true 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:19.792 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=75328 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 75328' 00:08:19.793 Process raid pid: 75328 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 75328 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 75328 ']' 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:19.793 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:19.793 18:58:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:19.793 [2024-12-05 18:58:37.210328] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:08:19.793 [2024-12-05 18:58:37.210551] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:20.053 [2024-12-05 18:58:37.365660] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:20.053 [2024-12-05 18:58:37.390477] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:20.053 [2024-12-05 18:58:37.433444] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:20.053 [2024-12-05 18:58:37.433559] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.624 [2024-12-05 18:58:38.032601] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:20.624 [2024-12-05 18:58:38.032657] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:20.624 [2024-12-05 18:58:38.032678] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:20.624 [2024-12-05 18:58:38.032688] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:20.624 [2024-12-05 18:58:38.032694] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:20.624 [2024-12-05 18:58:38.032705] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:20.624 "name": "Existed_Raid", 00:08:20.624 "uuid": "db53b61d-653d-41ce-8fdc-9988257325ef", 00:08:20.624 "strip_size_kb": 64, 00:08:20.624 "state": "configuring", 00:08:20.624 "raid_level": "raid0", 00:08:20.624 "superblock": true, 00:08:20.624 "num_base_bdevs": 3, 00:08:20.624 "num_base_bdevs_discovered": 0, 00:08:20.624 "num_base_bdevs_operational": 3, 00:08:20.624 "base_bdevs_list": [ 00:08:20.624 { 00:08:20.624 "name": "BaseBdev1", 00:08:20.624 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.624 "is_configured": false, 00:08:20.624 "data_offset": 0, 00:08:20.624 "data_size": 0 00:08:20.624 }, 00:08:20.624 { 00:08:20.624 "name": "BaseBdev2", 00:08:20.624 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.624 "is_configured": false, 00:08:20.624 "data_offset": 0, 00:08:20.624 "data_size": 0 00:08:20.624 }, 00:08:20.624 { 00:08:20.624 "name": "BaseBdev3", 00:08:20.624 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:20.624 "is_configured": false, 00:08:20.624 "data_offset": 0, 00:08:20.624 "data_size": 0 00:08:20.624 } 00:08:20.624 ] 00:08:20.624 }' 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:20.624 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.884 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:20.884 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.884 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:20.884 [2024-12-05 18:58:38.435768] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:20.884 [2024-12-05 18:58:38.435810] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:20.884 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:20.884 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:20.884 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:20.884 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.144 [2024-12-05 18:58:38.443812] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:21.144 [2024-12-05 18:58:38.443849] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:21.144 [2024-12-05 18:58:38.443858] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:21.144 [2024-12-05 18:58:38.443867] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:21.144 [2024-12-05 18:58:38.443873] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:21.144 [2024-12-05 18:58:38.443882] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.144 [2024-12-05 18:58:38.460797] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:21.144 BaseBdev1 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.144 [ 00:08:21.144 { 00:08:21.144 "name": "BaseBdev1", 00:08:21.144 "aliases": [ 00:08:21.144 "7c35481a-7148-47b6-80da-bcdd18b65462" 00:08:21.144 ], 00:08:21.144 "product_name": "Malloc disk", 00:08:21.144 "block_size": 512, 00:08:21.144 "num_blocks": 65536, 00:08:21.144 "uuid": "7c35481a-7148-47b6-80da-bcdd18b65462", 00:08:21.144 "assigned_rate_limits": { 00:08:21.144 "rw_ios_per_sec": 0, 00:08:21.144 "rw_mbytes_per_sec": 0, 00:08:21.144 "r_mbytes_per_sec": 0, 00:08:21.144 "w_mbytes_per_sec": 0 00:08:21.144 }, 00:08:21.144 "claimed": true, 00:08:21.144 "claim_type": "exclusive_write", 00:08:21.144 "zoned": false, 00:08:21.144 "supported_io_types": { 00:08:21.144 "read": true, 00:08:21.144 "write": true, 00:08:21.144 "unmap": true, 00:08:21.144 "flush": true, 00:08:21.144 "reset": true, 00:08:21.144 "nvme_admin": false, 00:08:21.144 "nvme_io": false, 00:08:21.144 "nvme_io_md": false, 00:08:21.144 "write_zeroes": true, 00:08:21.144 "zcopy": true, 00:08:21.144 "get_zone_info": false, 00:08:21.144 "zone_management": false, 00:08:21.144 "zone_append": false, 00:08:21.144 "compare": false, 00:08:21.144 "compare_and_write": false, 00:08:21.144 "abort": true, 00:08:21.144 "seek_hole": false, 00:08:21.144 "seek_data": false, 00:08:21.144 "copy": true, 00:08:21.144 "nvme_iov_md": false 00:08:21.144 }, 00:08:21.144 "memory_domains": [ 00:08:21.144 { 00:08:21.144 "dma_device_id": "system", 00:08:21.144 "dma_device_type": 1 00:08:21.144 }, 00:08:21.144 { 00:08:21.144 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:21.144 "dma_device_type": 2 00:08:21.144 } 00:08:21.144 ], 00:08:21.144 "driver_specific": {} 00:08:21.144 } 00:08:21.144 ] 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:21.144 "name": "Existed_Raid", 00:08:21.144 "uuid": "673c7738-b547-49c7-89a3-f752208bab77", 00:08:21.144 "strip_size_kb": 64, 00:08:21.144 "state": "configuring", 00:08:21.144 "raid_level": "raid0", 00:08:21.144 "superblock": true, 00:08:21.144 "num_base_bdevs": 3, 00:08:21.144 "num_base_bdevs_discovered": 1, 00:08:21.144 "num_base_bdevs_operational": 3, 00:08:21.144 "base_bdevs_list": [ 00:08:21.144 { 00:08:21.144 "name": "BaseBdev1", 00:08:21.144 "uuid": "7c35481a-7148-47b6-80da-bcdd18b65462", 00:08:21.144 "is_configured": true, 00:08:21.144 "data_offset": 2048, 00:08:21.144 "data_size": 63488 00:08:21.144 }, 00:08:21.144 { 00:08:21.144 "name": "BaseBdev2", 00:08:21.144 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:21.144 "is_configured": false, 00:08:21.144 "data_offset": 0, 00:08:21.144 "data_size": 0 00:08:21.144 }, 00:08:21.144 { 00:08:21.144 "name": "BaseBdev3", 00:08:21.144 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:21.144 "is_configured": false, 00:08:21.144 "data_offset": 0, 00:08:21.144 "data_size": 0 00:08:21.144 } 00:08:21.144 ] 00:08:21.144 }' 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:21.144 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.405 [2024-12-05 18:58:38.924060] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:21.405 [2024-12-05 18:58:38.924112] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.405 [2024-12-05 18:58:38.936077] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:21.405 [2024-12-05 18:58:38.937915] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:21.405 [2024-12-05 18:58:38.937960] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:21.405 [2024-12-05 18:58:38.937984] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:21.405 [2024-12-05 18:58:38.937994] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:21.405 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.665 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:21.665 "name": "Existed_Raid", 00:08:21.665 "uuid": "1aa54f44-25ff-4d53-8a04-2dcd71ce1643", 00:08:21.665 "strip_size_kb": 64, 00:08:21.665 "state": "configuring", 00:08:21.665 "raid_level": "raid0", 00:08:21.665 "superblock": true, 00:08:21.665 "num_base_bdevs": 3, 00:08:21.665 "num_base_bdevs_discovered": 1, 00:08:21.665 "num_base_bdevs_operational": 3, 00:08:21.665 "base_bdevs_list": [ 00:08:21.665 { 00:08:21.665 "name": "BaseBdev1", 00:08:21.665 "uuid": "7c35481a-7148-47b6-80da-bcdd18b65462", 00:08:21.665 "is_configured": true, 00:08:21.665 "data_offset": 2048, 00:08:21.665 "data_size": 63488 00:08:21.665 }, 00:08:21.665 { 00:08:21.665 "name": "BaseBdev2", 00:08:21.666 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:21.666 "is_configured": false, 00:08:21.666 "data_offset": 0, 00:08:21.666 "data_size": 0 00:08:21.666 }, 00:08:21.666 { 00:08:21.666 "name": "BaseBdev3", 00:08:21.666 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:21.666 "is_configured": false, 00:08:21.666 "data_offset": 0, 00:08:21.666 "data_size": 0 00:08:21.666 } 00:08:21.666 ] 00:08:21.666 }' 00:08:21.666 18:58:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:21.666 18:58:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.926 [2024-12-05 18:58:39.382600] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:21.926 BaseBdev2 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:21.926 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.927 [ 00:08:21.927 { 00:08:21.927 "name": "BaseBdev2", 00:08:21.927 "aliases": [ 00:08:21.927 "55db5530-5f83-40f5-8f1c-397c9d0d229b" 00:08:21.927 ], 00:08:21.927 "product_name": "Malloc disk", 00:08:21.927 "block_size": 512, 00:08:21.927 "num_blocks": 65536, 00:08:21.927 "uuid": "55db5530-5f83-40f5-8f1c-397c9d0d229b", 00:08:21.927 "assigned_rate_limits": { 00:08:21.927 "rw_ios_per_sec": 0, 00:08:21.927 "rw_mbytes_per_sec": 0, 00:08:21.927 "r_mbytes_per_sec": 0, 00:08:21.927 "w_mbytes_per_sec": 0 00:08:21.927 }, 00:08:21.927 "claimed": true, 00:08:21.927 "claim_type": "exclusive_write", 00:08:21.927 "zoned": false, 00:08:21.927 "supported_io_types": { 00:08:21.927 "read": true, 00:08:21.927 "write": true, 00:08:21.927 "unmap": true, 00:08:21.927 "flush": true, 00:08:21.927 "reset": true, 00:08:21.927 "nvme_admin": false, 00:08:21.927 "nvme_io": false, 00:08:21.927 "nvme_io_md": false, 00:08:21.927 "write_zeroes": true, 00:08:21.927 "zcopy": true, 00:08:21.927 "get_zone_info": false, 00:08:21.927 "zone_management": false, 00:08:21.927 "zone_append": false, 00:08:21.927 "compare": false, 00:08:21.927 "compare_and_write": false, 00:08:21.927 "abort": true, 00:08:21.927 "seek_hole": false, 00:08:21.927 "seek_data": false, 00:08:21.927 "copy": true, 00:08:21.927 "nvme_iov_md": false 00:08:21.927 }, 00:08:21.927 "memory_domains": [ 00:08:21.927 { 00:08:21.927 "dma_device_id": "system", 00:08:21.927 "dma_device_type": 1 00:08:21.927 }, 00:08:21.927 { 00:08:21.927 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:21.927 "dma_device_type": 2 00:08:21.927 } 00:08:21.927 ], 00:08:21.927 "driver_specific": {} 00:08:21.927 } 00:08:21.927 ] 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:21.927 "name": "Existed_Raid", 00:08:21.927 "uuid": "1aa54f44-25ff-4d53-8a04-2dcd71ce1643", 00:08:21.927 "strip_size_kb": 64, 00:08:21.927 "state": "configuring", 00:08:21.927 "raid_level": "raid0", 00:08:21.927 "superblock": true, 00:08:21.927 "num_base_bdevs": 3, 00:08:21.927 "num_base_bdevs_discovered": 2, 00:08:21.927 "num_base_bdevs_operational": 3, 00:08:21.927 "base_bdevs_list": [ 00:08:21.927 { 00:08:21.927 "name": "BaseBdev1", 00:08:21.927 "uuid": "7c35481a-7148-47b6-80da-bcdd18b65462", 00:08:21.927 "is_configured": true, 00:08:21.927 "data_offset": 2048, 00:08:21.927 "data_size": 63488 00:08:21.927 }, 00:08:21.927 { 00:08:21.927 "name": "BaseBdev2", 00:08:21.927 "uuid": "55db5530-5f83-40f5-8f1c-397c9d0d229b", 00:08:21.927 "is_configured": true, 00:08:21.927 "data_offset": 2048, 00:08:21.927 "data_size": 63488 00:08:21.927 }, 00:08:21.927 { 00:08:21.927 "name": "BaseBdev3", 00:08:21.927 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:21.927 "is_configured": false, 00:08:21.927 "data_offset": 0, 00:08:21.927 "data_size": 0 00:08:21.927 } 00:08:21.927 ] 00:08:21.927 }' 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:21.927 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.499 [2024-12-05 18:58:39.858625] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:22.499 [2024-12-05 18:58:39.858898] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:22.499 [2024-12-05 18:58:39.858931] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:22.499 BaseBdev3 00:08:22.499 [2024-12-05 18:58:39.859345] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:22.499 [2024-12-05 18:58:39.859557] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:22.499 [2024-12-05 18:58:39.859574] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:22.499 [2024-12-05 18:58:39.859800] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.499 [ 00:08:22.499 { 00:08:22.499 "name": "BaseBdev3", 00:08:22.499 "aliases": [ 00:08:22.499 "a2d86e7b-6a45-41de-94c1-129b893760eb" 00:08:22.499 ], 00:08:22.499 "product_name": "Malloc disk", 00:08:22.499 "block_size": 512, 00:08:22.499 "num_blocks": 65536, 00:08:22.499 "uuid": "a2d86e7b-6a45-41de-94c1-129b893760eb", 00:08:22.499 "assigned_rate_limits": { 00:08:22.499 "rw_ios_per_sec": 0, 00:08:22.499 "rw_mbytes_per_sec": 0, 00:08:22.499 "r_mbytes_per_sec": 0, 00:08:22.499 "w_mbytes_per_sec": 0 00:08:22.499 }, 00:08:22.499 "claimed": true, 00:08:22.499 "claim_type": "exclusive_write", 00:08:22.499 "zoned": false, 00:08:22.499 "supported_io_types": { 00:08:22.499 "read": true, 00:08:22.499 "write": true, 00:08:22.499 "unmap": true, 00:08:22.499 "flush": true, 00:08:22.499 "reset": true, 00:08:22.499 "nvme_admin": false, 00:08:22.499 "nvme_io": false, 00:08:22.499 "nvme_io_md": false, 00:08:22.499 "write_zeroes": true, 00:08:22.499 "zcopy": true, 00:08:22.499 "get_zone_info": false, 00:08:22.499 "zone_management": false, 00:08:22.499 "zone_append": false, 00:08:22.499 "compare": false, 00:08:22.499 "compare_and_write": false, 00:08:22.499 "abort": true, 00:08:22.499 "seek_hole": false, 00:08:22.499 "seek_data": false, 00:08:22.499 "copy": true, 00:08:22.499 "nvme_iov_md": false 00:08:22.499 }, 00:08:22.499 "memory_domains": [ 00:08:22.499 { 00:08:22.499 "dma_device_id": "system", 00:08:22.499 "dma_device_type": 1 00:08:22.499 }, 00:08:22.499 { 00:08:22.499 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:22.499 "dma_device_type": 2 00:08:22.499 } 00:08:22.499 ], 00:08:22.499 "driver_specific": {} 00:08:22.499 } 00:08:22.499 ] 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:22.499 "name": "Existed_Raid", 00:08:22.499 "uuid": "1aa54f44-25ff-4d53-8a04-2dcd71ce1643", 00:08:22.499 "strip_size_kb": 64, 00:08:22.499 "state": "online", 00:08:22.499 "raid_level": "raid0", 00:08:22.499 "superblock": true, 00:08:22.499 "num_base_bdevs": 3, 00:08:22.499 "num_base_bdevs_discovered": 3, 00:08:22.499 "num_base_bdevs_operational": 3, 00:08:22.499 "base_bdevs_list": [ 00:08:22.499 { 00:08:22.499 "name": "BaseBdev1", 00:08:22.499 "uuid": "7c35481a-7148-47b6-80da-bcdd18b65462", 00:08:22.499 "is_configured": true, 00:08:22.499 "data_offset": 2048, 00:08:22.499 "data_size": 63488 00:08:22.499 }, 00:08:22.499 { 00:08:22.499 "name": "BaseBdev2", 00:08:22.499 "uuid": "55db5530-5f83-40f5-8f1c-397c9d0d229b", 00:08:22.499 "is_configured": true, 00:08:22.499 "data_offset": 2048, 00:08:22.499 "data_size": 63488 00:08:22.499 }, 00:08:22.499 { 00:08:22.499 "name": "BaseBdev3", 00:08:22.499 "uuid": "a2d86e7b-6a45-41de-94c1-129b893760eb", 00:08:22.499 "is_configured": true, 00:08:22.499 "data_offset": 2048, 00:08:22.499 "data_size": 63488 00:08:22.499 } 00:08:22.499 ] 00:08:22.499 }' 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:22.499 18:58:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.080 [2024-12-05 18:58:40.334090] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:23.080 "name": "Existed_Raid", 00:08:23.080 "aliases": [ 00:08:23.080 "1aa54f44-25ff-4d53-8a04-2dcd71ce1643" 00:08:23.080 ], 00:08:23.080 "product_name": "Raid Volume", 00:08:23.080 "block_size": 512, 00:08:23.080 "num_blocks": 190464, 00:08:23.080 "uuid": "1aa54f44-25ff-4d53-8a04-2dcd71ce1643", 00:08:23.080 "assigned_rate_limits": { 00:08:23.080 "rw_ios_per_sec": 0, 00:08:23.080 "rw_mbytes_per_sec": 0, 00:08:23.080 "r_mbytes_per_sec": 0, 00:08:23.080 "w_mbytes_per_sec": 0 00:08:23.080 }, 00:08:23.080 "claimed": false, 00:08:23.080 "zoned": false, 00:08:23.080 "supported_io_types": { 00:08:23.080 "read": true, 00:08:23.080 "write": true, 00:08:23.080 "unmap": true, 00:08:23.080 "flush": true, 00:08:23.080 "reset": true, 00:08:23.080 "nvme_admin": false, 00:08:23.080 "nvme_io": false, 00:08:23.080 "nvme_io_md": false, 00:08:23.080 "write_zeroes": true, 00:08:23.080 "zcopy": false, 00:08:23.080 "get_zone_info": false, 00:08:23.080 "zone_management": false, 00:08:23.080 "zone_append": false, 00:08:23.080 "compare": false, 00:08:23.080 "compare_and_write": false, 00:08:23.080 "abort": false, 00:08:23.080 "seek_hole": false, 00:08:23.080 "seek_data": false, 00:08:23.080 "copy": false, 00:08:23.080 "nvme_iov_md": false 00:08:23.080 }, 00:08:23.080 "memory_domains": [ 00:08:23.080 { 00:08:23.080 "dma_device_id": "system", 00:08:23.080 "dma_device_type": 1 00:08:23.080 }, 00:08:23.080 { 00:08:23.080 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.080 "dma_device_type": 2 00:08:23.080 }, 00:08:23.080 { 00:08:23.080 "dma_device_id": "system", 00:08:23.080 "dma_device_type": 1 00:08:23.080 }, 00:08:23.080 { 00:08:23.080 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.080 "dma_device_type": 2 00:08:23.080 }, 00:08:23.080 { 00:08:23.080 "dma_device_id": "system", 00:08:23.080 "dma_device_type": 1 00:08:23.080 }, 00:08:23.080 { 00:08:23.080 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.080 "dma_device_type": 2 00:08:23.080 } 00:08:23.080 ], 00:08:23.080 "driver_specific": { 00:08:23.080 "raid": { 00:08:23.080 "uuid": "1aa54f44-25ff-4d53-8a04-2dcd71ce1643", 00:08:23.080 "strip_size_kb": 64, 00:08:23.080 "state": "online", 00:08:23.080 "raid_level": "raid0", 00:08:23.080 "superblock": true, 00:08:23.080 "num_base_bdevs": 3, 00:08:23.080 "num_base_bdevs_discovered": 3, 00:08:23.080 "num_base_bdevs_operational": 3, 00:08:23.080 "base_bdevs_list": [ 00:08:23.080 { 00:08:23.080 "name": "BaseBdev1", 00:08:23.080 "uuid": "7c35481a-7148-47b6-80da-bcdd18b65462", 00:08:23.080 "is_configured": true, 00:08:23.080 "data_offset": 2048, 00:08:23.080 "data_size": 63488 00:08:23.080 }, 00:08:23.080 { 00:08:23.080 "name": "BaseBdev2", 00:08:23.080 "uuid": "55db5530-5f83-40f5-8f1c-397c9d0d229b", 00:08:23.080 "is_configured": true, 00:08:23.080 "data_offset": 2048, 00:08:23.080 "data_size": 63488 00:08:23.080 }, 00:08:23.080 { 00:08:23.080 "name": "BaseBdev3", 00:08:23.080 "uuid": "a2d86e7b-6a45-41de-94c1-129b893760eb", 00:08:23.080 "is_configured": true, 00:08:23.080 "data_offset": 2048, 00:08:23.080 "data_size": 63488 00:08:23.080 } 00:08:23.080 ] 00:08:23.080 } 00:08:23.080 } 00:08:23.080 }' 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:23.080 BaseBdev2 00:08:23.080 BaseBdev3' 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.080 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.081 [2024-12-05 18:58:40.601402] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:23.081 [2024-12-05 18:58:40.601430] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:23.081 [2024-12-05 18:58:40.601485] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.081 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.342 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.342 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:23.342 "name": "Existed_Raid", 00:08:23.342 "uuid": "1aa54f44-25ff-4d53-8a04-2dcd71ce1643", 00:08:23.342 "strip_size_kb": 64, 00:08:23.342 "state": "offline", 00:08:23.342 "raid_level": "raid0", 00:08:23.342 "superblock": true, 00:08:23.342 "num_base_bdevs": 3, 00:08:23.342 "num_base_bdevs_discovered": 2, 00:08:23.342 "num_base_bdevs_operational": 2, 00:08:23.342 "base_bdevs_list": [ 00:08:23.342 { 00:08:23.342 "name": null, 00:08:23.342 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.342 "is_configured": false, 00:08:23.342 "data_offset": 0, 00:08:23.342 "data_size": 63488 00:08:23.342 }, 00:08:23.342 { 00:08:23.342 "name": "BaseBdev2", 00:08:23.342 "uuid": "55db5530-5f83-40f5-8f1c-397c9d0d229b", 00:08:23.342 "is_configured": true, 00:08:23.342 "data_offset": 2048, 00:08:23.342 "data_size": 63488 00:08:23.342 }, 00:08:23.342 { 00:08:23.342 "name": "BaseBdev3", 00:08:23.342 "uuid": "a2d86e7b-6a45-41de-94c1-129b893760eb", 00:08:23.342 "is_configured": true, 00:08:23.342 "data_offset": 2048, 00:08:23.342 "data_size": 63488 00:08:23.342 } 00:08:23.342 ] 00:08:23.342 }' 00:08:23.342 18:58:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:23.342 18:58:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.602 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:23.602 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:23.602 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.602 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.603 [2024-12-05 18:58:41.099934] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.603 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.865 [2024-12-05 18:58:41.166952] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:23.865 [2024-12-05 18:58:41.167004] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.865 BaseBdev2 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.865 [ 00:08:23.865 { 00:08:23.865 "name": "BaseBdev2", 00:08:23.865 "aliases": [ 00:08:23.865 "23dea4ef-696d-4fc0-ae20-a76b8a699173" 00:08:23.865 ], 00:08:23.865 "product_name": "Malloc disk", 00:08:23.865 "block_size": 512, 00:08:23.865 "num_blocks": 65536, 00:08:23.865 "uuid": "23dea4ef-696d-4fc0-ae20-a76b8a699173", 00:08:23.865 "assigned_rate_limits": { 00:08:23.865 "rw_ios_per_sec": 0, 00:08:23.865 "rw_mbytes_per_sec": 0, 00:08:23.865 "r_mbytes_per_sec": 0, 00:08:23.865 "w_mbytes_per_sec": 0 00:08:23.865 }, 00:08:23.865 "claimed": false, 00:08:23.865 "zoned": false, 00:08:23.865 "supported_io_types": { 00:08:23.865 "read": true, 00:08:23.865 "write": true, 00:08:23.865 "unmap": true, 00:08:23.865 "flush": true, 00:08:23.865 "reset": true, 00:08:23.865 "nvme_admin": false, 00:08:23.865 "nvme_io": false, 00:08:23.865 "nvme_io_md": false, 00:08:23.865 "write_zeroes": true, 00:08:23.865 "zcopy": true, 00:08:23.865 "get_zone_info": false, 00:08:23.865 "zone_management": false, 00:08:23.865 "zone_append": false, 00:08:23.865 "compare": false, 00:08:23.865 "compare_and_write": false, 00:08:23.865 "abort": true, 00:08:23.865 "seek_hole": false, 00:08:23.865 "seek_data": false, 00:08:23.865 "copy": true, 00:08:23.865 "nvme_iov_md": false 00:08:23.865 }, 00:08:23.865 "memory_domains": [ 00:08:23.865 { 00:08:23.865 "dma_device_id": "system", 00:08:23.865 "dma_device_type": 1 00:08:23.865 }, 00:08:23.865 { 00:08:23.865 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.865 "dma_device_type": 2 00:08:23.865 } 00:08:23.865 ], 00:08:23.865 "driver_specific": {} 00:08:23.865 } 00:08:23.865 ] 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.865 BaseBdev3 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.865 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.865 [ 00:08:23.865 { 00:08:23.865 "name": "BaseBdev3", 00:08:23.865 "aliases": [ 00:08:23.865 "9f2aa0a6-9164-4b5d-a752-2d72da1ad373" 00:08:23.865 ], 00:08:23.865 "product_name": "Malloc disk", 00:08:23.865 "block_size": 512, 00:08:23.865 "num_blocks": 65536, 00:08:23.865 "uuid": "9f2aa0a6-9164-4b5d-a752-2d72da1ad373", 00:08:23.865 "assigned_rate_limits": { 00:08:23.865 "rw_ios_per_sec": 0, 00:08:23.865 "rw_mbytes_per_sec": 0, 00:08:23.865 "r_mbytes_per_sec": 0, 00:08:23.865 "w_mbytes_per_sec": 0 00:08:23.865 }, 00:08:23.865 "claimed": false, 00:08:23.865 "zoned": false, 00:08:23.865 "supported_io_types": { 00:08:23.865 "read": true, 00:08:23.865 "write": true, 00:08:23.865 "unmap": true, 00:08:23.865 "flush": true, 00:08:23.865 "reset": true, 00:08:23.865 "nvme_admin": false, 00:08:23.865 "nvme_io": false, 00:08:23.865 "nvme_io_md": false, 00:08:23.865 "write_zeroes": true, 00:08:23.865 "zcopy": true, 00:08:23.865 "get_zone_info": false, 00:08:23.865 "zone_management": false, 00:08:23.865 "zone_append": false, 00:08:23.865 "compare": false, 00:08:23.865 "compare_and_write": false, 00:08:23.865 "abort": true, 00:08:23.865 "seek_hole": false, 00:08:23.865 "seek_data": false, 00:08:23.865 "copy": true, 00:08:23.866 "nvme_iov_md": false 00:08:23.866 }, 00:08:23.866 "memory_domains": [ 00:08:23.866 { 00:08:23.866 "dma_device_id": "system", 00:08:23.866 "dma_device_type": 1 00:08:23.866 }, 00:08:23.866 { 00:08:23.866 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.866 "dma_device_type": 2 00:08:23.866 } 00:08:23.866 ], 00:08:23.866 "driver_specific": {} 00:08:23.866 } 00:08:23.866 ] 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.866 [2024-12-05 18:58:41.326339] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:23.866 [2024-12-05 18:58:41.326378] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:23.866 [2024-12-05 18:58:41.326398] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:23.866 [2024-12-05 18:58:41.328363] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:23.866 "name": "Existed_Raid", 00:08:23.866 "uuid": "063f1fae-bfd2-43ac-9c1c-6628be36e416", 00:08:23.866 "strip_size_kb": 64, 00:08:23.866 "state": "configuring", 00:08:23.866 "raid_level": "raid0", 00:08:23.866 "superblock": true, 00:08:23.866 "num_base_bdevs": 3, 00:08:23.866 "num_base_bdevs_discovered": 2, 00:08:23.866 "num_base_bdevs_operational": 3, 00:08:23.866 "base_bdevs_list": [ 00:08:23.866 { 00:08:23.866 "name": "BaseBdev1", 00:08:23.866 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.866 "is_configured": false, 00:08:23.866 "data_offset": 0, 00:08:23.866 "data_size": 0 00:08:23.866 }, 00:08:23.866 { 00:08:23.866 "name": "BaseBdev2", 00:08:23.866 "uuid": "23dea4ef-696d-4fc0-ae20-a76b8a699173", 00:08:23.866 "is_configured": true, 00:08:23.866 "data_offset": 2048, 00:08:23.866 "data_size": 63488 00:08:23.866 }, 00:08:23.866 { 00:08:23.866 "name": "BaseBdev3", 00:08:23.866 "uuid": "9f2aa0a6-9164-4b5d-a752-2d72da1ad373", 00:08:23.866 "is_configured": true, 00:08:23.866 "data_offset": 2048, 00:08:23.866 "data_size": 63488 00:08:23.866 } 00:08:23.866 ] 00:08:23.866 }' 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:23.866 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.438 [2024-12-05 18:58:41.781536] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:24.438 "name": "Existed_Raid", 00:08:24.438 "uuid": "063f1fae-bfd2-43ac-9c1c-6628be36e416", 00:08:24.438 "strip_size_kb": 64, 00:08:24.438 "state": "configuring", 00:08:24.438 "raid_level": "raid0", 00:08:24.438 "superblock": true, 00:08:24.438 "num_base_bdevs": 3, 00:08:24.438 "num_base_bdevs_discovered": 1, 00:08:24.438 "num_base_bdevs_operational": 3, 00:08:24.438 "base_bdevs_list": [ 00:08:24.438 { 00:08:24.438 "name": "BaseBdev1", 00:08:24.438 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:24.438 "is_configured": false, 00:08:24.438 "data_offset": 0, 00:08:24.438 "data_size": 0 00:08:24.438 }, 00:08:24.438 { 00:08:24.438 "name": null, 00:08:24.438 "uuid": "23dea4ef-696d-4fc0-ae20-a76b8a699173", 00:08:24.438 "is_configured": false, 00:08:24.438 "data_offset": 0, 00:08:24.438 "data_size": 63488 00:08:24.438 }, 00:08:24.438 { 00:08:24.438 "name": "BaseBdev3", 00:08:24.438 "uuid": "9f2aa0a6-9164-4b5d-a752-2d72da1ad373", 00:08:24.438 "is_configured": true, 00:08:24.438 "data_offset": 2048, 00:08:24.438 "data_size": 63488 00:08:24.438 } 00:08:24.438 ] 00:08:24.438 }' 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:24.438 18:58:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.699 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:24.699 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.699 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.699 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:24.699 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.959 [2024-12-05 18:58:42.267773] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:24.959 BaseBdev1 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.959 [ 00:08:24.959 { 00:08:24.959 "name": "BaseBdev1", 00:08:24.959 "aliases": [ 00:08:24.959 "a0bf4c92-1785-4522-840d-f206b0b133e9" 00:08:24.959 ], 00:08:24.959 "product_name": "Malloc disk", 00:08:24.959 "block_size": 512, 00:08:24.959 "num_blocks": 65536, 00:08:24.959 "uuid": "a0bf4c92-1785-4522-840d-f206b0b133e9", 00:08:24.959 "assigned_rate_limits": { 00:08:24.959 "rw_ios_per_sec": 0, 00:08:24.959 "rw_mbytes_per_sec": 0, 00:08:24.959 "r_mbytes_per_sec": 0, 00:08:24.959 "w_mbytes_per_sec": 0 00:08:24.959 }, 00:08:24.959 "claimed": true, 00:08:24.959 "claim_type": "exclusive_write", 00:08:24.959 "zoned": false, 00:08:24.959 "supported_io_types": { 00:08:24.959 "read": true, 00:08:24.959 "write": true, 00:08:24.959 "unmap": true, 00:08:24.959 "flush": true, 00:08:24.959 "reset": true, 00:08:24.959 "nvme_admin": false, 00:08:24.959 "nvme_io": false, 00:08:24.959 "nvme_io_md": false, 00:08:24.959 "write_zeroes": true, 00:08:24.959 "zcopy": true, 00:08:24.959 "get_zone_info": false, 00:08:24.959 "zone_management": false, 00:08:24.959 "zone_append": false, 00:08:24.959 "compare": false, 00:08:24.959 "compare_and_write": false, 00:08:24.959 "abort": true, 00:08:24.959 "seek_hole": false, 00:08:24.959 "seek_data": false, 00:08:24.959 "copy": true, 00:08:24.959 "nvme_iov_md": false 00:08:24.959 }, 00:08:24.959 "memory_domains": [ 00:08:24.959 { 00:08:24.959 "dma_device_id": "system", 00:08:24.959 "dma_device_type": 1 00:08:24.959 }, 00:08:24.959 { 00:08:24.959 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:24.959 "dma_device_type": 2 00:08:24.959 } 00:08:24.959 ], 00:08:24.959 "driver_specific": {} 00:08:24.959 } 00:08:24.959 ] 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:24.959 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:24.959 "name": "Existed_Raid", 00:08:24.959 "uuid": "063f1fae-bfd2-43ac-9c1c-6628be36e416", 00:08:24.959 "strip_size_kb": 64, 00:08:24.959 "state": "configuring", 00:08:24.959 "raid_level": "raid0", 00:08:24.959 "superblock": true, 00:08:24.959 "num_base_bdevs": 3, 00:08:24.959 "num_base_bdevs_discovered": 2, 00:08:24.959 "num_base_bdevs_operational": 3, 00:08:24.959 "base_bdevs_list": [ 00:08:24.959 { 00:08:24.959 "name": "BaseBdev1", 00:08:24.959 "uuid": "a0bf4c92-1785-4522-840d-f206b0b133e9", 00:08:24.959 "is_configured": true, 00:08:24.959 "data_offset": 2048, 00:08:24.959 "data_size": 63488 00:08:24.959 }, 00:08:24.959 { 00:08:24.959 "name": null, 00:08:24.959 "uuid": "23dea4ef-696d-4fc0-ae20-a76b8a699173", 00:08:24.959 "is_configured": false, 00:08:24.959 "data_offset": 0, 00:08:24.960 "data_size": 63488 00:08:24.960 }, 00:08:24.960 { 00:08:24.960 "name": "BaseBdev3", 00:08:24.960 "uuid": "9f2aa0a6-9164-4b5d-a752-2d72da1ad373", 00:08:24.960 "is_configured": true, 00:08:24.960 "data_offset": 2048, 00:08:24.960 "data_size": 63488 00:08:24.960 } 00:08:24.960 ] 00:08:24.960 }' 00:08:24.960 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:24.960 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.220 [2024-12-05 18:58:42.651497] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:25.220 "name": "Existed_Raid", 00:08:25.220 "uuid": "063f1fae-bfd2-43ac-9c1c-6628be36e416", 00:08:25.220 "strip_size_kb": 64, 00:08:25.220 "state": "configuring", 00:08:25.220 "raid_level": "raid0", 00:08:25.220 "superblock": true, 00:08:25.220 "num_base_bdevs": 3, 00:08:25.220 "num_base_bdevs_discovered": 1, 00:08:25.220 "num_base_bdevs_operational": 3, 00:08:25.220 "base_bdevs_list": [ 00:08:25.220 { 00:08:25.220 "name": "BaseBdev1", 00:08:25.220 "uuid": "a0bf4c92-1785-4522-840d-f206b0b133e9", 00:08:25.220 "is_configured": true, 00:08:25.220 "data_offset": 2048, 00:08:25.220 "data_size": 63488 00:08:25.220 }, 00:08:25.220 { 00:08:25.220 "name": null, 00:08:25.220 "uuid": "23dea4ef-696d-4fc0-ae20-a76b8a699173", 00:08:25.220 "is_configured": false, 00:08:25.220 "data_offset": 0, 00:08:25.220 "data_size": 63488 00:08:25.220 }, 00:08:25.220 { 00:08:25.220 "name": null, 00:08:25.220 "uuid": "9f2aa0a6-9164-4b5d-a752-2d72da1ad373", 00:08:25.220 "is_configured": false, 00:08:25.220 "data_offset": 0, 00:08:25.220 "data_size": 63488 00:08:25.220 } 00:08:25.220 ] 00:08:25.220 }' 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:25.220 18:58:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.790 [2024-12-05 18:58:43.130694] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:25.790 "name": "Existed_Raid", 00:08:25.790 "uuid": "063f1fae-bfd2-43ac-9c1c-6628be36e416", 00:08:25.790 "strip_size_kb": 64, 00:08:25.790 "state": "configuring", 00:08:25.790 "raid_level": "raid0", 00:08:25.790 "superblock": true, 00:08:25.790 "num_base_bdevs": 3, 00:08:25.790 "num_base_bdevs_discovered": 2, 00:08:25.790 "num_base_bdevs_operational": 3, 00:08:25.790 "base_bdevs_list": [ 00:08:25.790 { 00:08:25.790 "name": "BaseBdev1", 00:08:25.790 "uuid": "a0bf4c92-1785-4522-840d-f206b0b133e9", 00:08:25.790 "is_configured": true, 00:08:25.790 "data_offset": 2048, 00:08:25.790 "data_size": 63488 00:08:25.790 }, 00:08:25.790 { 00:08:25.790 "name": null, 00:08:25.790 "uuid": "23dea4ef-696d-4fc0-ae20-a76b8a699173", 00:08:25.790 "is_configured": false, 00:08:25.790 "data_offset": 0, 00:08:25.790 "data_size": 63488 00:08:25.790 }, 00:08:25.790 { 00:08:25.790 "name": "BaseBdev3", 00:08:25.790 "uuid": "9f2aa0a6-9164-4b5d-a752-2d72da1ad373", 00:08:25.790 "is_configured": true, 00:08:25.790 "data_offset": 2048, 00:08:25.790 "data_size": 63488 00:08:25.790 } 00:08:25.790 ] 00:08:25.790 }' 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:25.790 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.050 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.050 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.050 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.050 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:26.050 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.309 [2024-12-05 18:58:43.633819] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:26.309 "name": "Existed_Raid", 00:08:26.309 "uuid": "063f1fae-bfd2-43ac-9c1c-6628be36e416", 00:08:26.309 "strip_size_kb": 64, 00:08:26.309 "state": "configuring", 00:08:26.309 "raid_level": "raid0", 00:08:26.309 "superblock": true, 00:08:26.309 "num_base_bdevs": 3, 00:08:26.309 "num_base_bdevs_discovered": 1, 00:08:26.309 "num_base_bdevs_operational": 3, 00:08:26.309 "base_bdevs_list": [ 00:08:26.309 { 00:08:26.309 "name": null, 00:08:26.309 "uuid": "a0bf4c92-1785-4522-840d-f206b0b133e9", 00:08:26.309 "is_configured": false, 00:08:26.309 "data_offset": 0, 00:08:26.309 "data_size": 63488 00:08:26.309 }, 00:08:26.309 { 00:08:26.309 "name": null, 00:08:26.309 "uuid": "23dea4ef-696d-4fc0-ae20-a76b8a699173", 00:08:26.309 "is_configured": false, 00:08:26.309 "data_offset": 0, 00:08:26.309 "data_size": 63488 00:08:26.309 }, 00:08:26.309 { 00:08:26.309 "name": "BaseBdev3", 00:08:26.309 "uuid": "9f2aa0a6-9164-4b5d-a752-2d72da1ad373", 00:08:26.309 "is_configured": true, 00:08:26.309 "data_offset": 2048, 00:08:26.309 "data_size": 63488 00:08:26.309 } 00:08:26.309 ] 00:08:26.309 }' 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:26.309 18:58:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.569 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:26.569 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.569 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.569 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.569 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.569 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:26.569 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:26.569 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.569 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.569 [2024-12-05 18:58:44.119787] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:26.829 "name": "Existed_Raid", 00:08:26.829 "uuid": "063f1fae-bfd2-43ac-9c1c-6628be36e416", 00:08:26.829 "strip_size_kb": 64, 00:08:26.829 "state": "configuring", 00:08:26.829 "raid_level": "raid0", 00:08:26.829 "superblock": true, 00:08:26.829 "num_base_bdevs": 3, 00:08:26.829 "num_base_bdevs_discovered": 2, 00:08:26.829 "num_base_bdevs_operational": 3, 00:08:26.829 "base_bdevs_list": [ 00:08:26.829 { 00:08:26.829 "name": null, 00:08:26.829 "uuid": "a0bf4c92-1785-4522-840d-f206b0b133e9", 00:08:26.829 "is_configured": false, 00:08:26.829 "data_offset": 0, 00:08:26.829 "data_size": 63488 00:08:26.829 }, 00:08:26.829 { 00:08:26.829 "name": "BaseBdev2", 00:08:26.829 "uuid": "23dea4ef-696d-4fc0-ae20-a76b8a699173", 00:08:26.829 "is_configured": true, 00:08:26.829 "data_offset": 2048, 00:08:26.829 "data_size": 63488 00:08:26.829 }, 00:08:26.829 { 00:08:26.829 "name": "BaseBdev3", 00:08:26.829 "uuid": "9f2aa0a6-9164-4b5d-a752-2d72da1ad373", 00:08:26.829 "is_configured": true, 00:08:26.829 "data_offset": 2048, 00:08:26.829 "data_size": 63488 00:08:26.829 } 00:08:26.829 ] 00:08:26.829 }' 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:26.829 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u a0bf4c92-1785-4522-840d-f206b0b133e9 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.089 [2024-12-05 18:58:44.618776] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:27.089 [2024-12-05 18:58:44.618938] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:27.089 [2024-12-05 18:58:44.618977] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:27.089 [2024-12-05 18:58:44.619213] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:27.089 [2024-12-05 18:58:44.619342] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:27.089 [2024-12-05 18:58:44.619356] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:27.089 [2024-12-05 18:58:44.619461] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:27.089 NewBaseBdev 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.089 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.349 [ 00:08:27.349 { 00:08:27.349 "name": "NewBaseBdev", 00:08:27.349 "aliases": [ 00:08:27.349 "a0bf4c92-1785-4522-840d-f206b0b133e9" 00:08:27.349 ], 00:08:27.349 "product_name": "Malloc disk", 00:08:27.349 "block_size": 512, 00:08:27.349 "num_blocks": 65536, 00:08:27.349 "uuid": "a0bf4c92-1785-4522-840d-f206b0b133e9", 00:08:27.349 "assigned_rate_limits": { 00:08:27.349 "rw_ios_per_sec": 0, 00:08:27.349 "rw_mbytes_per_sec": 0, 00:08:27.349 "r_mbytes_per_sec": 0, 00:08:27.349 "w_mbytes_per_sec": 0 00:08:27.349 }, 00:08:27.349 "claimed": true, 00:08:27.349 "claim_type": "exclusive_write", 00:08:27.349 "zoned": false, 00:08:27.349 "supported_io_types": { 00:08:27.349 "read": true, 00:08:27.349 "write": true, 00:08:27.349 "unmap": true, 00:08:27.349 "flush": true, 00:08:27.349 "reset": true, 00:08:27.349 "nvme_admin": false, 00:08:27.349 "nvme_io": false, 00:08:27.349 "nvme_io_md": false, 00:08:27.349 "write_zeroes": true, 00:08:27.349 "zcopy": true, 00:08:27.349 "get_zone_info": false, 00:08:27.349 "zone_management": false, 00:08:27.349 "zone_append": false, 00:08:27.349 "compare": false, 00:08:27.349 "compare_and_write": false, 00:08:27.349 "abort": true, 00:08:27.349 "seek_hole": false, 00:08:27.349 "seek_data": false, 00:08:27.349 "copy": true, 00:08:27.349 "nvme_iov_md": false 00:08:27.349 }, 00:08:27.349 "memory_domains": [ 00:08:27.349 { 00:08:27.349 "dma_device_id": "system", 00:08:27.349 "dma_device_type": 1 00:08:27.349 }, 00:08:27.349 { 00:08:27.349 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:27.349 "dma_device_type": 2 00:08:27.349 } 00:08:27.349 ], 00:08:27.349 "driver_specific": {} 00:08:27.349 } 00:08:27.349 ] 00:08:27.349 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.349 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:27.349 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:27.349 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:27.350 "name": "Existed_Raid", 00:08:27.350 "uuid": "063f1fae-bfd2-43ac-9c1c-6628be36e416", 00:08:27.350 "strip_size_kb": 64, 00:08:27.350 "state": "online", 00:08:27.350 "raid_level": "raid0", 00:08:27.350 "superblock": true, 00:08:27.350 "num_base_bdevs": 3, 00:08:27.350 "num_base_bdevs_discovered": 3, 00:08:27.350 "num_base_bdevs_operational": 3, 00:08:27.350 "base_bdevs_list": [ 00:08:27.350 { 00:08:27.350 "name": "NewBaseBdev", 00:08:27.350 "uuid": "a0bf4c92-1785-4522-840d-f206b0b133e9", 00:08:27.350 "is_configured": true, 00:08:27.350 "data_offset": 2048, 00:08:27.350 "data_size": 63488 00:08:27.350 }, 00:08:27.350 { 00:08:27.350 "name": "BaseBdev2", 00:08:27.350 "uuid": "23dea4ef-696d-4fc0-ae20-a76b8a699173", 00:08:27.350 "is_configured": true, 00:08:27.350 "data_offset": 2048, 00:08:27.350 "data_size": 63488 00:08:27.350 }, 00:08:27.350 { 00:08:27.350 "name": "BaseBdev3", 00:08:27.350 "uuid": "9f2aa0a6-9164-4b5d-a752-2d72da1ad373", 00:08:27.350 "is_configured": true, 00:08:27.350 "data_offset": 2048, 00:08:27.350 "data_size": 63488 00:08:27.350 } 00:08:27.350 ] 00:08:27.350 }' 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:27.350 18:58:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.610 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:27.610 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:27.610 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:27.610 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:27.610 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:27.610 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:27.610 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:27.610 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.610 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.610 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:27.610 [2024-12-05 18:58:45.066331] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:27.610 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.610 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:27.610 "name": "Existed_Raid", 00:08:27.610 "aliases": [ 00:08:27.610 "063f1fae-bfd2-43ac-9c1c-6628be36e416" 00:08:27.610 ], 00:08:27.610 "product_name": "Raid Volume", 00:08:27.610 "block_size": 512, 00:08:27.610 "num_blocks": 190464, 00:08:27.610 "uuid": "063f1fae-bfd2-43ac-9c1c-6628be36e416", 00:08:27.610 "assigned_rate_limits": { 00:08:27.610 "rw_ios_per_sec": 0, 00:08:27.610 "rw_mbytes_per_sec": 0, 00:08:27.610 "r_mbytes_per_sec": 0, 00:08:27.610 "w_mbytes_per_sec": 0 00:08:27.610 }, 00:08:27.610 "claimed": false, 00:08:27.610 "zoned": false, 00:08:27.610 "supported_io_types": { 00:08:27.610 "read": true, 00:08:27.610 "write": true, 00:08:27.610 "unmap": true, 00:08:27.610 "flush": true, 00:08:27.610 "reset": true, 00:08:27.610 "nvme_admin": false, 00:08:27.610 "nvme_io": false, 00:08:27.610 "nvme_io_md": false, 00:08:27.610 "write_zeroes": true, 00:08:27.610 "zcopy": false, 00:08:27.610 "get_zone_info": false, 00:08:27.610 "zone_management": false, 00:08:27.610 "zone_append": false, 00:08:27.610 "compare": false, 00:08:27.610 "compare_and_write": false, 00:08:27.610 "abort": false, 00:08:27.610 "seek_hole": false, 00:08:27.610 "seek_data": false, 00:08:27.610 "copy": false, 00:08:27.610 "nvme_iov_md": false 00:08:27.610 }, 00:08:27.610 "memory_domains": [ 00:08:27.610 { 00:08:27.610 "dma_device_id": "system", 00:08:27.610 "dma_device_type": 1 00:08:27.610 }, 00:08:27.610 { 00:08:27.610 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:27.610 "dma_device_type": 2 00:08:27.610 }, 00:08:27.610 { 00:08:27.610 "dma_device_id": "system", 00:08:27.610 "dma_device_type": 1 00:08:27.610 }, 00:08:27.610 { 00:08:27.610 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:27.610 "dma_device_type": 2 00:08:27.610 }, 00:08:27.610 { 00:08:27.610 "dma_device_id": "system", 00:08:27.610 "dma_device_type": 1 00:08:27.610 }, 00:08:27.610 { 00:08:27.610 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:27.610 "dma_device_type": 2 00:08:27.610 } 00:08:27.610 ], 00:08:27.610 "driver_specific": { 00:08:27.610 "raid": { 00:08:27.610 "uuid": "063f1fae-bfd2-43ac-9c1c-6628be36e416", 00:08:27.610 "strip_size_kb": 64, 00:08:27.610 "state": "online", 00:08:27.610 "raid_level": "raid0", 00:08:27.610 "superblock": true, 00:08:27.610 "num_base_bdevs": 3, 00:08:27.610 "num_base_bdevs_discovered": 3, 00:08:27.610 "num_base_bdevs_operational": 3, 00:08:27.610 "base_bdevs_list": [ 00:08:27.610 { 00:08:27.610 "name": "NewBaseBdev", 00:08:27.610 "uuid": "a0bf4c92-1785-4522-840d-f206b0b133e9", 00:08:27.610 "is_configured": true, 00:08:27.610 "data_offset": 2048, 00:08:27.610 "data_size": 63488 00:08:27.610 }, 00:08:27.611 { 00:08:27.611 "name": "BaseBdev2", 00:08:27.611 "uuid": "23dea4ef-696d-4fc0-ae20-a76b8a699173", 00:08:27.611 "is_configured": true, 00:08:27.611 "data_offset": 2048, 00:08:27.611 "data_size": 63488 00:08:27.611 }, 00:08:27.611 { 00:08:27.611 "name": "BaseBdev3", 00:08:27.611 "uuid": "9f2aa0a6-9164-4b5d-a752-2d72da1ad373", 00:08:27.611 "is_configured": true, 00:08:27.611 "data_offset": 2048, 00:08:27.611 "data_size": 63488 00:08:27.611 } 00:08:27.611 ] 00:08:27.611 } 00:08:27.611 } 00:08:27.611 }' 00:08:27.611 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:27.611 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:27.611 BaseBdev2 00:08:27.611 BaseBdev3' 00:08:27.611 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:27.871 [2024-12-05 18:58:45.305680] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:27.871 [2024-12-05 18:58:45.305707] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:27.871 [2024-12-05 18:58:45.305777] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:27.871 [2024-12-05 18:58:45.305828] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:27.871 [2024-12-05 18:58:45.305851] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 75328 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 75328 ']' 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 75328 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 75328 00:08:27.871 killing process with pid 75328 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 75328' 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 75328 00:08:27.871 [2024-12-05 18:58:45.351424] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:27.871 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 75328 00:08:27.871 [2024-12-05 18:58:45.381601] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:28.131 18:58:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:28.131 00:08:28.131 real 0m8.471s 00:08:28.131 user 0m14.457s 00:08:28.131 sys 0m1.686s 00:08:28.131 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:28.131 18:58:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:28.131 ************************************ 00:08:28.131 END TEST raid_state_function_test_sb 00:08:28.131 ************************************ 00:08:28.131 18:58:45 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 3 00:08:28.131 18:58:45 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:08:28.131 18:58:45 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:28.131 18:58:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:28.131 ************************************ 00:08:28.131 START TEST raid_superblock_test 00:08:28.131 ************************************ 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 3 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=75926 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 75926 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 75926 ']' 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:28.131 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:28.131 18:58:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.391 [2024-12-05 18:58:45.742123] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:08:28.391 [2024-12-05 18:58:45.742277] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid75926 ] 00:08:28.391 [2024-12-05 18:58:45.895734] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:28.391 [2024-12-05 18:58:45.920370] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:28.650 [2024-12-05 18:58:45.963441] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:28.650 [2024-12-05 18:58:45.963486] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.220 malloc1 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.220 [2024-12-05 18:58:46.587638] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:29.220 [2024-12-05 18:58:46.587725] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:29.220 [2024-12-05 18:58:46.587762] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:08:29.220 [2024-12-05 18:58:46.587776] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:29.220 [2024-12-05 18:58:46.589841] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:29.220 [2024-12-05 18:58:46.589879] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:29.220 pt1 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.220 malloc2 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.220 [2024-12-05 18:58:46.616337] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:29.220 [2024-12-05 18:58:46.616395] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:29.220 [2024-12-05 18:58:46.616412] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:29.220 [2024-12-05 18:58:46.616422] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:29.220 [2024-12-05 18:58:46.618555] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:29.220 [2024-12-05 18:58:46.618594] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:29.220 pt2 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.220 malloc3 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.220 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.220 [2024-12-05 18:58:46.645015] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:29.220 [2024-12-05 18:58:46.645070] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:29.220 [2024-12-05 18:58:46.645105] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:29.220 [2024-12-05 18:58:46.645116] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:29.220 [2024-12-05 18:58:46.647272] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:29.220 [2024-12-05 18:58:46.647323] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:29.220 pt3 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.221 [2024-12-05 18:58:46.657063] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:29.221 [2024-12-05 18:58:46.658900] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:29.221 [2024-12-05 18:58:46.658995] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:29.221 [2024-12-05 18:58:46.659145] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:29.221 [2024-12-05 18:58:46.659157] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:29.221 [2024-12-05 18:58:46.659422] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:29.221 [2024-12-05 18:58:46.659574] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:29.221 [2024-12-05 18:58:46.659603] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:08:29.221 [2024-12-05 18:58:46.659732] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:29.221 "name": "raid_bdev1", 00:08:29.221 "uuid": "ab649339-0e16-4bb5-9563-36f4ce0268a9", 00:08:29.221 "strip_size_kb": 64, 00:08:29.221 "state": "online", 00:08:29.221 "raid_level": "raid0", 00:08:29.221 "superblock": true, 00:08:29.221 "num_base_bdevs": 3, 00:08:29.221 "num_base_bdevs_discovered": 3, 00:08:29.221 "num_base_bdevs_operational": 3, 00:08:29.221 "base_bdevs_list": [ 00:08:29.221 { 00:08:29.221 "name": "pt1", 00:08:29.221 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:29.221 "is_configured": true, 00:08:29.221 "data_offset": 2048, 00:08:29.221 "data_size": 63488 00:08:29.221 }, 00:08:29.221 { 00:08:29.221 "name": "pt2", 00:08:29.221 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:29.221 "is_configured": true, 00:08:29.221 "data_offset": 2048, 00:08:29.221 "data_size": 63488 00:08:29.221 }, 00:08:29.221 { 00:08:29.221 "name": "pt3", 00:08:29.221 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:29.221 "is_configured": true, 00:08:29.221 "data_offset": 2048, 00:08:29.221 "data_size": 63488 00:08:29.221 } 00:08:29.221 ] 00:08:29.221 }' 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:29.221 18:58:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.791 [2024-12-05 18:58:47.096532] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:29.791 "name": "raid_bdev1", 00:08:29.791 "aliases": [ 00:08:29.791 "ab649339-0e16-4bb5-9563-36f4ce0268a9" 00:08:29.791 ], 00:08:29.791 "product_name": "Raid Volume", 00:08:29.791 "block_size": 512, 00:08:29.791 "num_blocks": 190464, 00:08:29.791 "uuid": "ab649339-0e16-4bb5-9563-36f4ce0268a9", 00:08:29.791 "assigned_rate_limits": { 00:08:29.791 "rw_ios_per_sec": 0, 00:08:29.791 "rw_mbytes_per_sec": 0, 00:08:29.791 "r_mbytes_per_sec": 0, 00:08:29.791 "w_mbytes_per_sec": 0 00:08:29.791 }, 00:08:29.791 "claimed": false, 00:08:29.791 "zoned": false, 00:08:29.791 "supported_io_types": { 00:08:29.791 "read": true, 00:08:29.791 "write": true, 00:08:29.791 "unmap": true, 00:08:29.791 "flush": true, 00:08:29.791 "reset": true, 00:08:29.791 "nvme_admin": false, 00:08:29.791 "nvme_io": false, 00:08:29.791 "nvme_io_md": false, 00:08:29.791 "write_zeroes": true, 00:08:29.791 "zcopy": false, 00:08:29.791 "get_zone_info": false, 00:08:29.791 "zone_management": false, 00:08:29.791 "zone_append": false, 00:08:29.791 "compare": false, 00:08:29.791 "compare_and_write": false, 00:08:29.791 "abort": false, 00:08:29.791 "seek_hole": false, 00:08:29.791 "seek_data": false, 00:08:29.791 "copy": false, 00:08:29.791 "nvme_iov_md": false 00:08:29.791 }, 00:08:29.791 "memory_domains": [ 00:08:29.791 { 00:08:29.791 "dma_device_id": "system", 00:08:29.791 "dma_device_type": 1 00:08:29.791 }, 00:08:29.791 { 00:08:29.791 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.791 "dma_device_type": 2 00:08:29.791 }, 00:08:29.791 { 00:08:29.791 "dma_device_id": "system", 00:08:29.791 "dma_device_type": 1 00:08:29.791 }, 00:08:29.791 { 00:08:29.791 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.791 "dma_device_type": 2 00:08:29.791 }, 00:08:29.791 { 00:08:29.791 "dma_device_id": "system", 00:08:29.791 "dma_device_type": 1 00:08:29.791 }, 00:08:29.791 { 00:08:29.791 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.791 "dma_device_type": 2 00:08:29.791 } 00:08:29.791 ], 00:08:29.791 "driver_specific": { 00:08:29.791 "raid": { 00:08:29.791 "uuid": "ab649339-0e16-4bb5-9563-36f4ce0268a9", 00:08:29.791 "strip_size_kb": 64, 00:08:29.791 "state": "online", 00:08:29.791 "raid_level": "raid0", 00:08:29.791 "superblock": true, 00:08:29.791 "num_base_bdevs": 3, 00:08:29.791 "num_base_bdevs_discovered": 3, 00:08:29.791 "num_base_bdevs_operational": 3, 00:08:29.791 "base_bdevs_list": [ 00:08:29.791 { 00:08:29.791 "name": "pt1", 00:08:29.791 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:29.791 "is_configured": true, 00:08:29.791 "data_offset": 2048, 00:08:29.791 "data_size": 63488 00:08:29.791 }, 00:08:29.791 { 00:08:29.791 "name": "pt2", 00:08:29.791 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:29.791 "is_configured": true, 00:08:29.791 "data_offset": 2048, 00:08:29.791 "data_size": 63488 00:08:29.791 }, 00:08:29.791 { 00:08:29.791 "name": "pt3", 00:08:29.791 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:29.791 "is_configured": true, 00:08:29.791 "data_offset": 2048, 00:08:29.791 "data_size": 63488 00:08:29.791 } 00:08:29.791 ] 00:08:29.791 } 00:08:29.791 } 00:08:29.791 }' 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:29.791 pt2 00:08:29.791 pt3' 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:29.791 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:29.792 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.792 [2024-12-05 18:58:47.348038] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:30.051 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.051 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=ab649339-0e16-4bb5-9563-36f4ce0268a9 00:08:30.051 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z ab649339-0e16-4bb5-9563-36f4ce0268a9 ']' 00:08:30.051 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:30.051 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.051 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.051 [2024-12-05 18:58:47.379787] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:30.051 [2024-12-05 18:58:47.379819] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:30.051 [2024-12-05 18:58:47.379899] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:30.051 [2024-12-05 18:58:47.379957] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:30.051 [2024-12-05 18:58:47.379970] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:08:30.051 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.051 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:30.051 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.051 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.051 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.051 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.052 [2024-12-05 18:58:47.511810] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:30.052 [2024-12-05 18:58:47.513689] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:30.052 [2024-12-05 18:58:47.513736] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:08:30.052 [2024-12-05 18:58:47.513780] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:30.052 [2024-12-05 18:58:47.513824] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:30.052 [2024-12-05 18:58:47.513852] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:08:30.052 [2024-12-05 18:58:47.513865] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:30.052 [2024-12-05 18:58:47.513875] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:08:30.052 request: 00:08:30.052 { 00:08:30.052 "name": "raid_bdev1", 00:08:30.052 "raid_level": "raid0", 00:08:30.052 "base_bdevs": [ 00:08:30.052 "malloc1", 00:08:30.052 "malloc2", 00:08:30.052 "malloc3" 00:08:30.052 ], 00:08:30.052 "strip_size_kb": 64, 00:08:30.052 "superblock": false, 00:08:30.052 "method": "bdev_raid_create", 00:08:30.052 "req_id": 1 00:08:30.052 } 00:08:30.052 Got JSON-RPC error response 00:08:30.052 response: 00:08:30.052 { 00:08:30.052 "code": -17, 00:08:30.052 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:30.052 } 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.052 [2024-12-05 18:58:47.579774] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:30.052 [2024-12-05 18:58:47.579868] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:30.052 [2024-12-05 18:58:47.579900] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:30.052 [2024-12-05 18:58:47.579928] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:30.052 [2024-12-05 18:58:47.581992] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:30.052 [2024-12-05 18:58:47.582065] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:30.052 [2024-12-05 18:58:47.582143] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:30.052 [2024-12-05 18:58:47.582203] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:30.052 pt1 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:30.052 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:30.053 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:30.053 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:30.053 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.053 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:30.053 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.053 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.053 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.313 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:30.313 "name": "raid_bdev1", 00:08:30.313 "uuid": "ab649339-0e16-4bb5-9563-36f4ce0268a9", 00:08:30.313 "strip_size_kb": 64, 00:08:30.313 "state": "configuring", 00:08:30.313 "raid_level": "raid0", 00:08:30.313 "superblock": true, 00:08:30.313 "num_base_bdevs": 3, 00:08:30.313 "num_base_bdevs_discovered": 1, 00:08:30.313 "num_base_bdevs_operational": 3, 00:08:30.313 "base_bdevs_list": [ 00:08:30.313 { 00:08:30.313 "name": "pt1", 00:08:30.313 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:30.313 "is_configured": true, 00:08:30.313 "data_offset": 2048, 00:08:30.313 "data_size": 63488 00:08:30.313 }, 00:08:30.313 { 00:08:30.313 "name": null, 00:08:30.313 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:30.313 "is_configured": false, 00:08:30.313 "data_offset": 2048, 00:08:30.313 "data_size": 63488 00:08:30.313 }, 00:08:30.313 { 00:08:30.313 "name": null, 00:08:30.313 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:30.313 "is_configured": false, 00:08:30.313 "data_offset": 2048, 00:08:30.313 "data_size": 63488 00:08:30.313 } 00:08:30.313 ] 00:08:30.313 }' 00:08:30.313 18:58:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:30.313 18:58:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.573 [2024-12-05 18:58:48.023804] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:30.573 [2024-12-05 18:58:48.023905] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:30.573 [2024-12-05 18:58:48.023942] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:08:30.573 [2024-12-05 18:58:48.023973] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:30.573 [2024-12-05 18:58:48.024342] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:30.573 [2024-12-05 18:58:48.024399] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:30.573 [2024-12-05 18:58:48.024487] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:30.573 [2024-12-05 18:58:48.024536] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:30.573 pt2 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.573 [2024-12-05 18:58:48.035817] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:30.573 "name": "raid_bdev1", 00:08:30.573 "uuid": "ab649339-0e16-4bb5-9563-36f4ce0268a9", 00:08:30.573 "strip_size_kb": 64, 00:08:30.573 "state": "configuring", 00:08:30.573 "raid_level": "raid0", 00:08:30.573 "superblock": true, 00:08:30.573 "num_base_bdevs": 3, 00:08:30.573 "num_base_bdevs_discovered": 1, 00:08:30.573 "num_base_bdevs_operational": 3, 00:08:30.573 "base_bdevs_list": [ 00:08:30.573 { 00:08:30.573 "name": "pt1", 00:08:30.573 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:30.573 "is_configured": true, 00:08:30.573 "data_offset": 2048, 00:08:30.573 "data_size": 63488 00:08:30.573 }, 00:08:30.573 { 00:08:30.573 "name": null, 00:08:30.573 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:30.573 "is_configured": false, 00:08:30.573 "data_offset": 0, 00:08:30.573 "data_size": 63488 00:08:30.573 }, 00:08:30.573 { 00:08:30.573 "name": null, 00:08:30.573 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:30.573 "is_configured": false, 00:08:30.573 "data_offset": 2048, 00:08:30.573 "data_size": 63488 00:08:30.573 } 00:08:30.573 ] 00:08:30.573 }' 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:30.573 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.144 [2024-12-05 18:58:48.475769] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:31.144 [2024-12-05 18:58:48.475879] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:31.144 [2024-12-05 18:58:48.475901] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:31.144 [2024-12-05 18:58:48.475910] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:31.144 [2024-12-05 18:58:48.476278] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:31.144 [2024-12-05 18:58:48.476295] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:31.144 [2024-12-05 18:58:48.476360] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:31.144 [2024-12-05 18:58:48.476379] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:31.144 pt2 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.144 [2024-12-05 18:58:48.487781] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:31.144 [2024-12-05 18:58:48.487825] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:31.144 [2024-12-05 18:58:48.487841] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:08:31.144 [2024-12-05 18:58:48.487849] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:31.144 [2024-12-05 18:58:48.488156] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:31.144 [2024-12-05 18:58:48.488172] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:31.144 [2024-12-05 18:58:48.488222] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:31.144 [2024-12-05 18:58:48.488238] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:31.144 [2024-12-05 18:58:48.488323] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:31.144 [2024-12-05 18:58:48.488331] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:31.144 [2024-12-05 18:58:48.488547] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:31.144 [2024-12-05 18:58:48.488655] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:31.144 [2024-12-05 18:58:48.488695] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:31.144 [2024-12-05 18:58:48.488792] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:31.144 pt3 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.144 "name": "raid_bdev1", 00:08:31.144 "uuid": "ab649339-0e16-4bb5-9563-36f4ce0268a9", 00:08:31.144 "strip_size_kb": 64, 00:08:31.144 "state": "online", 00:08:31.144 "raid_level": "raid0", 00:08:31.144 "superblock": true, 00:08:31.144 "num_base_bdevs": 3, 00:08:31.144 "num_base_bdevs_discovered": 3, 00:08:31.144 "num_base_bdevs_operational": 3, 00:08:31.144 "base_bdevs_list": [ 00:08:31.144 { 00:08:31.144 "name": "pt1", 00:08:31.144 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:31.144 "is_configured": true, 00:08:31.144 "data_offset": 2048, 00:08:31.144 "data_size": 63488 00:08:31.144 }, 00:08:31.144 { 00:08:31.144 "name": "pt2", 00:08:31.144 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:31.144 "is_configured": true, 00:08:31.144 "data_offset": 2048, 00:08:31.144 "data_size": 63488 00:08:31.144 }, 00:08:31.144 { 00:08:31.144 "name": "pt3", 00:08:31.144 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:31.144 "is_configured": true, 00:08:31.144 "data_offset": 2048, 00:08:31.144 "data_size": 63488 00:08:31.144 } 00:08:31.144 ] 00:08:31.144 }' 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.144 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.403 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:31.403 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:31.403 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:31.403 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:31.403 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:31.403 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:31.403 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:31.403 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:31.403 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.403 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.403 [2024-12-05 18:58:48.960056] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:31.662 18:58:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.662 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:31.662 "name": "raid_bdev1", 00:08:31.662 "aliases": [ 00:08:31.662 "ab649339-0e16-4bb5-9563-36f4ce0268a9" 00:08:31.662 ], 00:08:31.662 "product_name": "Raid Volume", 00:08:31.662 "block_size": 512, 00:08:31.662 "num_blocks": 190464, 00:08:31.662 "uuid": "ab649339-0e16-4bb5-9563-36f4ce0268a9", 00:08:31.662 "assigned_rate_limits": { 00:08:31.662 "rw_ios_per_sec": 0, 00:08:31.662 "rw_mbytes_per_sec": 0, 00:08:31.662 "r_mbytes_per_sec": 0, 00:08:31.662 "w_mbytes_per_sec": 0 00:08:31.662 }, 00:08:31.662 "claimed": false, 00:08:31.662 "zoned": false, 00:08:31.662 "supported_io_types": { 00:08:31.662 "read": true, 00:08:31.662 "write": true, 00:08:31.662 "unmap": true, 00:08:31.662 "flush": true, 00:08:31.662 "reset": true, 00:08:31.663 "nvme_admin": false, 00:08:31.663 "nvme_io": false, 00:08:31.663 "nvme_io_md": false, 00:08:31.663 "write_zeroes": true, 00:08:31.663 "zcopy": false, 00:08:31.663 "get_zone_info": false, 00:08:31.663 "zone_management": false, 00:08:31.663 "zone_append": false, 00:08:31.663 "compare": false, 00:08:31.663 "compare_and_write": false, 00:08:31.663 "abort": false, 00:08:31.663 "seek_hole": false, 00:08:31.663 "seek_data": false, 00:08:31.663 "copy": false, 00:08:31.663 "nvme_iov_md": false 00:08:31.663 }, 00:08:31.663 "memory_domains": [ 00:08:31.663 { 00:08:31.663 "dma_device_id": "system", 00:08:31.663 "dma_device_type": 1 00:08:31.663 }, 00:08:31.663 { 00:08:31.663 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:31.663 "dma_device_type": 2 00:08:31.663 }, 00:08:31.663 { 00:08:31.663 "dma_device_id": "system", 00:08:31.663 "dma_device_type": 1 00:08:31.663 }, 00:08:31.663 { 00:08:31.663 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:31.663 "dma_device_type": 2 00:08:31.663 }, 00:08:31.663 { 00:08:31.663 "dma_device_id": "system", 00:08:31.663 "dma_device_type": 1 00:08:31.663 }, 00:08:31.663 { 00:08:31.663 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:31.663 "dma_device_type": 2 00:08:31.663 } 00:08:31.663 ], 00:08:31.663 "driver_specific": { 00:08:31.663 "raid": { 00:08:31.663 "uuid": "ab649339-0e16-4bb5-9563-36f4ce0268a9", 00:08:31.663 "strip_size_kb": 64, 00:08:31.663 "state": "online", 00:08:31.663 "raid_level": "raid0", 00:08:31.663 "superblock": true, 00:08:31.663 "num_base_bdevs": 3, 00:08:31.663 "num_base_bdevs_discovered": 3, 00:08:31.663 "num_base_bdevs_operational": 3, 00:08:31.663 "base_bdevs_list": [ 00:08:31.663 { 00:08:31.663 "name": "pt1", 00:08:31.663 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:31.663 "is_configured": true, 00:08:31.663 "data_offset": 2048, 00:08:31.663 "data_size": 63488 00:08:31.663 }, 00:08:31.663 { 00:08:31.663 "name": "pt2", 00:08:31.663 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:31.663 "is_configured": true, 00:08:31.663 "data_offset": 2048, 00:08:31.663 "data_size": 63488 00:08:31.663 }, 00:08:31.663 { 00:08:31.663 "name": "pt3", 00:08:31.663 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:31.663 "is_configured": true, 00:08:31.663 "data_offset": 2048, 00:08:31.663 "data_size": 63488 00:08:31.663 } 00:08:31.663 ] 00:08:31.663 } 00:08:31.663 } 00:08:31.663 }' 00:08:31.663 18:58:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:31.663 pt2 00:08:31.663 pt3' 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:31.663 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.922 [2024-12-05 18:58:49.228043] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' ab649339-0e16-4bb5-9563-36f4ce0268a9 '!=' ab649339-0e16-4bb5-9563-36f4ce0268a9 ']' 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 75926 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 75926 ']' 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 75926 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 75926 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 75926' 00:08:31.922 killing process with pid 75926 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 75926 00:08:31.922 [2024-12-05 18:58:49.300934] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:31.922 [2024-12-05 18:58:49.301011] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:31.922 [2024-12-05 18:58:49.301085] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:31.922 [2024-12-05 18:58:49.301093] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:31.922 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 75926 00:08:31.922 [2024-12-05 18:58:49.333612] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:32.182 18:58:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:32.182 00:08:32.182 real 0m3.888s 00:08:32.182 user 0m6.155s 00:08:32.182 sys 0m0.830s 00:08:32.182 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:32.182 18:58:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.182 ************************************ 00:08:32.182 END TEST raid_superblock_test 00:08:32.182 ************************************ 00:08:32.182 18:58:49 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 3 read 00:08:32.182 18:58:49 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:32.182 18:58:49 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:32.182 18:58:49 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:32.182 ************************************ 00:08:32.182 START TEST raid_read_error_test 00:08:32.182 ************************************ 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 3 read 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.Rc0M3RiGDW 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76163 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76163 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 76163 ']' 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:32.182 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:32.182 18:58:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.182 [2024-12-05 18:58:49.725614] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:08:32.182 [2024-12-05 18:58:49.725849] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76163 ] 00:08:32.442 [2024-12-05 18:58:49.879573] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:32.442 [2024-12-05 18:58:49.904856] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:32.442 [2024-12-05 18:58:49.947822] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:32.442 [2024-12-05 18:58:49.947860] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:33.012 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:33.012 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:33.012 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:33.012 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:33.012 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.012 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.012 BaseBdev1_malloc 00:08:33.012 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.012 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:33.012 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.012 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.012 true 00:08:33.012 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.012 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:33.012 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.012 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.012 [2024-12-05 18:58:50.564830] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:33.012 [2024-12-05 18:58:50.564884] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:33.012 [2024-12-05 18:58:50.564904] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:33.012 [2024-12-05 18:58:50.564920] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:33.012 [2024-12-05 18:58:50.566995] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:33.012 [2024-12-05 18:58:50.567036] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:33.271 BaseBdev1 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.271 BaseBdev2_malloc 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.271 true 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.271 [2024-12-05 18:58:50.605374] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:33.271 [2024-12-05 18:58:50.605420] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:33.271 [2024-12-05 18:58:50.605453] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:33.271 [2024-12-05 18:58:50.605469] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:33.271 [2024-12-05 18:58:50.607477] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:33.271 [2024-12-05 18:58:50.607516] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:33.271 BaseBdev2 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.271 BaseBdev3_malloc 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.271 true 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.271 [2024-12-05 18:58:50.645843] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:33.271 [2024-12-05 18:58:50.645941] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:33.271 [2024-12-05 18:58:50.645978] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:33.271 [2024-12-05 18:58:50.646004] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:33.271 [2024-12-05 18:58:50.648024] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:33.271 [2024-12-05 18:58:50.648095] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:33.271 BaseBdev3 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.271 [2024-12-05 18:58:50.657879] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:33.271 [2024-12-05 18:58:50.659530] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:33.271 [2024-12-05 18:58:50.659593] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:33.271 [2024-12-05 18:58:50.659795] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:33.271 [2024-12-05 18:58:50.659810] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:33.271 [2024-12-05 18:58:50.660075] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:08:33.271 [2024-12-05 18:58:50.660230] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:33.271 [2024-12-05 18:58:50.660247] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:33.271 [2024-12-05 18:58:50.660369] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:33.271 "name": "raid_bdev1", 00:08:33.271 "uuid": "4e9c94cc-4308-42e6-83ec-b6ee5db64f5a", 00:08:33.271 "strip_size_kb": 64, 00:08:33.271 "state": "online", 00:08:33.271 "raid_level": "raid0", 00:08:33.271 "superblock": true, 00:08:33.271 "num_base_bdevs": 3, 00:08:33.271 "num_base_bdevs_discovered": 3, 00:08:33.271 "num_base_bdevs_operational": 3, 00:08:33.271 "base_bdevs_list": [ 00:08:33.271 { 00:08:33.271 "name": "BaseBdev1", 00:08:33.271 "uuid": "43486bb3-338c-5d28-b47a-e35e09e71e94", 00:08:33.271 "is_configured": true, 00:08:33.271 "data_offset": 2048, 00:08:33.271 "data_size": 63488 00:08:33.271 }, 00:08:33.271 { 00:08:33.271 "name": "BaseBdev2", 00:08:33.271 "uuid": "a7aec855-58ad-5870-9e73-5d4ec25ada76", 00:08:33.271 "is_configured": true, 00:08:33.271 "data_offset": 2048, 00:08:33.271 "data_size": 63488 00:08:33.271 }, 00:08:33.271 { 00:08:33.271 "name": "BaseBdev3", 00:08:33.271 "uuid": "896483a2-2610-5a5b-a990-e277683864ff", 00:08:33.271 "is_configured": true, 00:08:33.271 "data_offset": 2048, 00:08:33.271 "data_size": 63488 00:08:33.271 } 00:08:33.271 ] 00:08:33.271 }' 00:08:33.271 18:58:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:33.272 18:58:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.839 18:58:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:33.839 18:58:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:33.839 [2024-12-05 18:58:51.217321] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:34.776 "name": "raid_bdev1", 00:08:34.776 "uuid": "4e9c94cc-4308-42e6-83ec-b6ee5db64f5a", 00:08:34.776 "strip_size_kb": 64, 00:08:34.776 "state": "online", 00:08:34.776 "raid_level": "raid0", 00:08:34.776 "superblock": true, 00:08:34.776 "num_base_bdevs": 3, 00:08:34.776 "num_base_bdevs_discovered": 3, 00:08:34.776 "num_base_bdevs_operational": 3, 00:08:34.776 "base_bdevs_list": [ 00:08:34.776 { 00:08:34.776 "name": "BaseBdev1", 00:08:34.776 "uuid": "43486bb3-338c-5d28-b47a-e35e09e71e94", 00:08:34.776 "is_configured": true, 00:08:34.776 "data_offset": 2048, 00:08:34.776 "data_size": 63488 00:08:34.776 }, 00:08:34.776 { 00:08:34.776 "name": "BaseBdev2", 00:08:34.776 "uuid": "a7aec855-58ad-5870-9e73-5d4ec25ada76", 00:08:34.776 "is_configured": true, 00:08:34.776 "data_offset": 2048, 00:08:34.776 "data_size": 63488 00:08:34.776 }, 00:08:34.776 { 00:08:34.776 "name": "BaseBdev3", 00:08:34.776 "uuid": "896483a2-2610-5a5b-a990-e277683864ff", 00:08:34.776 "is_configured": true, 00:08:34.776 "data_offset": 2048, 00:08:34.776 "data_size": 63488 00:08:34.776 } 00:08:34.776 ] 00:08:34.776 }' 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:34.776 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.346 [2024-12-05 18:58:52.609136] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:35.346 [2024-12-05 18:58:52.609241] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:35.346 [2024-12-05 18:58:52.611757] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:35.346 [2024-12-05 18:58:52.611849] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:35.346 [2024-12-05 18:58:52.611901] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:35.346 [2024-12-05 18:58:52.611955] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:35.346 { 00:08:35.346 "results": [ 00:08:35.346 { 00:08:35.346 "job": "raid_bdev1", 00:08:35.346 "core_mask": "0x1", 00:08:35.346 "workload": "randrw", 00:08:35.346 "percentage": 50, 00:08:35.346 "status": "finished", 00:08:35.346 "queue_depth": 1, 00:08:35.346 "io_size": 131072, 00:08:35.346 "runtime": 1.392871, 00:08:35.346 "iops": 17093.470967519606, 00:08:35.346 "mibps": 2136.683870939951, 00:08:35.346 "io_failed": 1, 00:08:35.346 "io_timeout": 0, 00:08:35.346 "avg_latency_us": 80.7836302313255, 00:08:35.346 "min_latency_us": 19.563318777292576, 00:08:35.346 "max_latency_us": 1359.3711790393013 00:08:35.346 } 00:08:35.346 ], 00:08:35.346 "core_count": 1 00:08:35.346 } 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76163 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 76163 ']' 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 76163 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76163 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76163' 00:08:35.346 killing process with pid 76163 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 76163 00:08:35.346 [2024-12-05 18:58:52.661350] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 76163 00:08:35.346 [2024-12-05 18:58:52.685951] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.Rc0M3RiGDW 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:08:35.346 00:08:35.346 real 0m3.276s 00:08:35.346 user 0m4.194s 00:08:35.346 sys 0m0.518s 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:35.346 18:58:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.346 ************************************ 00:08:35.346 END TEST raid_read_error_test 00:08:35.346 ************************************ 00:08:35.606 18:58:52 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 3 write 00:08:35.606 18:58:52 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:35.606 18:58:52 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:35.606 18:58:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:35.606 ************************************ 00:08:35.606 START TEST raid_write_error_test 00:08:35.606 ************************************ 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 3 write 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.LsMhk9wZr9 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76292 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76292 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 76292 ']' 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:35.606 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:35.606 18:58:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:35.606 [2024-12-05 18:58:53.074349] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:08:35.606 [2024-12-05 18:58:53.074470] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76292 ] 00:08:35.873 [2024-12-05 18:58:53.227060] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:35.873 [2024-12-05 18:58:53.252360] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:35.873 [2024-12-05 18:58:53.295625] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:35.873 [2024-12-05 18:58:53.295684] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.445 BaseBdev1_malloc 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.445 true 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.445 [2024-12-05 18:58:53.923601] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:36.445 [2024-12-05 18:58:53.923653] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:36.445 [2024-12-05 18:58:53.923708] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:36.445 [2024-12-05 18:58:53.923718] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:36.445 [2024-12-05 18:58:53.925840] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:36.445 [2024-12-05 18:58:53.925921] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:36.445 BaseBdev1 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.445 BaseBdev2_malloc 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.445 true 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.445 [2024-12-05 18:58:53.964232] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:36.445 [2024-12-05 18:58:53.964280] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:36.445 [2024-12-05 18:58:53.964297] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:36.445 [2024-12-05 18:58:53.964314] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:36.445 [2024-12-05 18:58:53.966377] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:36.445 [2024-12-05 18:58:53.966417] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:36.445 BaseBdev2 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.445 BaseBdev3_malloc 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.445 true 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.445 18:58:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.703 [2024-12-05 18:58:54.004849] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:36.703 [2024-12-05 18:58:54.004892] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:36.703 [2024-12-05 18:58:54.004926] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:36.703 [2024-12-05 18:58:54.004935] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:36.703 [2024-12-05 18:58:54.006922] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:36.703 [2024-12-05 18:58:54.006959] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:36.703 BaseBdev3 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.703 [2024-12-05 18:58:54.016878] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:36.703 [2024-12-05 18:58:54.018630] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:36.703 [2024-12-05 18:58:54.018715] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:36.703 [2024-12-05 18:58:54.018886] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:36.703 [2024-12-05 18:58:54.018905] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:36.703 [2024-12-05 18:58:54.019150] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:08:36.703 [2024-12-05 18:58:54.019299] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:36.703 [2024-12-05 18:58:54.019314] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:36.703 [2024-12-05 18:58:54.019431] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:36.703 "name": "raid_bdev1", 00:08:36.703 "uuid": "dab772e1-30ea-4f38-8c40-5298272d0a1a", 00:08:36.703 "strip_size_kb": 64, 00:08:36.703 "state": "online", 00:08:36.703 "raid_level": "raid0", 00:08:36.703 "superblock": true, 00:08:36.703 "num_base_bdevs": 3, 00:08:36.703 "num_base_bdevs_discovered": 3, 00:08:36.703 "num_base_bdevs_operational": 3, 00:08:36.703 "base_bdevs_list": [ 00:08:36.703 { 00:08:36.703 "name": "BaseBdev1", 00:08:36.703 "uuid": "aa7f6c7f-f32b-5fb6-815a-edc735407346", 00:08:36.703 "is_configured": true, 00:08:36.703 "data_offset": 2048, 00:08:36.703 "data_size": 63488 00:08:36.703 }, 00:08:36.703 { 00:08:36.703 "name": "BaseBdev2", 00:08:36.703 "uuid": "9b3ca53f-f1f6-52ec-93c2-90f0a5d2d141", 00:08:36.703 "is_configured": true, 00:08:36.703 "data_offset": 2048, 00:08:36.703 "data_size": 63488 00:08:36.703 }, 00:08:36.703 { 00:08:36.703 "name": "BaseBdev3", 00:08:36.703 "uuid": "6f020473-d174-5d8c-8ea7-3b7a9ee6cafa", 00:08:36.703 "is_configured": true, 00:08:36.703 "data_offset": 2048, 00:08:36.703 "data_size": 63488 00:08:36.703 } 00:08:36.703 ] 00:08:36.703 }' 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:36.703 18:58:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:36.960 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:36.960 18:58:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:37.218 [2024-12-05 18:58:54.528406] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:08:38.151 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:38.152 "name": "raid_bdev1", 00:08:38.152 "uuid": "dab772e1-30ea-4f38-8c40-5298272d0a1a", 00:08:38.152 "strip_size_kb": 64, 00:08:38.152 "state": "online", 00:08:38.152 "raid_level": "raid0", 00:08:38.152 "superblock": true, 00:08:38.152 "num_base_bdevs": 3, 00:08:38.152 "num_base_bdevs_discovered": 3, 00:08:38.152 "num_base_bdevs_operational": 3, 00:08:38.152 "base_bdevs_list": [ 00:08:38.152 { 00:08:38.152 "name": "BaseBdev1", 00:08:38.152 "uuid": "aa7f6c7f-f32b-5fb6-815a-edc735407346", 00:08:38.152 "is_configured": true, 00:08:38.152 "data_offset": 2048, 00:08:38.152 "data_size": 63488 00:08:38.152 }, 00:08:38.152 { 00:08:38.152 "name": "BaseBdev2", 00:08:38.152 "uuid": "9b3ca53f-f1f6-52ec-93c2-90f0a5d2d141", 00:08:38.152 "is_configured": true, 00:08:38.152 "data_offset": 2048, 00:08:38.152 "data_size": 63488 00:08:38.152 }, 00:08:38.152 { 00:08:38.152 "name": "BaseBdev3", 00:08:38.152 "uuid": "6f020473-d174-5d8c-8ea7-3b7a9ee6cafa", 00:08:38.152 "is_configured": true, 00:08:38.152 "data_offset": 2048, 00:08:38.152 "data_size": 63488 00:08:38.152 } 00:08:38.152 ] 00:08:38.152 }' 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:38.152 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.409 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:38.409 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:38.409 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.409 [2024-12-05 18:58:55.936304] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:38.409 [2024-12-05 18:58:55.936399] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:38.409 [2024-12-05 18:58:55.938951] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:38.409 [2024-12-05 18:58:55.939041] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:38.409 [2024-12-05 18:58:55.939092] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:38.409 [2024-12-05 18:58:55.939133] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:38.409 { 00:08:38.409 "results": [ 00:08:38.409 { 00:08:38.409 "job": "raid_bdev1", 00:08:38.409 "core_mask": "0x1", 00:08:38.409 "workload": "randrw", 00:08:38.409 "percentage": 50, 00:08:38.409 "status": "finished", 00:08:38.409 "queue_depth": 1, 00:08:38.409 "io_size": 131072, 00:08:38.409 "runtime": 1.40903, 00:08:38.409 "iops": 17099.706890555914, 00:08:38.409 "mibps": 2137.463361319489, 00:08:38.409 "io_failed": 1, 00:08:38.409 "io_timeout": 0, 00:08:38.409 "avg_latency_us": 80.77549278646842, 00:08:38.409 "min_latency_us": 25.041048034934498, 00:08:38.409 "max_latency_us": 1380.8349344978167 00:08:38.409 } 00:08:38.409 ], 00:08:38.409 "core_count": 1 00:08:38.409 } 00:08:38.409 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:38.409 18:58:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76292 00:08:38.409 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 76292 ']' 00:08:38.409 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 76292 00:08:38.409 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:08:38.409 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:38.409 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76292 00:08:38.667 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:38.667 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:38.667 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76292' 00:08:38.667 killing process with pid 76292 00:08:38.667 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 76292 00:08:38.667 [2024-12-05 18:58:55.988398] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:38.667 18:58:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 76292 00:08:38.667 [2024-12-05 18:58:56.013617] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:38.667 18:58:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.LsMhk9wZr9 00:08:38.667 18:58:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:38.667 18:58:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:38.925 18:58:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.71 00:08:38.925 18:58:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:38.925 18:58:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:38.925 18:58:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:38.925 18:58:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.71 != \0\.\0\0 ]] 00:08:38.925 00:08:38.925 real 0m3.254s 00:08:38.925 user 0m4.143s 00:08:38.925 sys 0m0.515s 00:08:38.925 18:58:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:38.925 18:58:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.925 ************************************ 00:08:38.925 END TEST raid_write_error_test 00:08:38.925 ************************************ 00:08:38.925 18:58:56 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:38.925 18:58:56 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 3 false 00:08:38.925 18:58:56 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:38.925 18:58:56 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:38.925 18:58:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:38.925 ************************************ 00:08:38.925 START TEST raid_state_function_test 00:08:38.925 ************************************ 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 3 false 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:38.925 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:38.926 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:38.926 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:38.926 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:38.926 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:38.926 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:38.926 Process raid pid: 76426 00:08:38.926 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=76426 00:08:38.926 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:38.926 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 76426' 00:08:38.926 18:58:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 76426 00:08:38.926 18:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 76426 ']' 00:08:38.926 18:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:38.926 18:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:38.926 18:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:38.926 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:38.926 18:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:38.926 18:58:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:38.926 [2024-12-05 18:58:56.395157] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:08:38.926 [2024-12-05 18:58:56.395700] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:39.184 [2024-12-05 18:58:56.550413] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:39.184 [2024-12-05 18:58:56.575346] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:39.184 [2024-12-05 18:58:56.618364] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:39.184 [2024-12-05 18:58:56.618398] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:39.751 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:39.751 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:08:39.751 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.752 [2024-12-05 18:58:57.209403] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:39.752 [2024-12-05 18:58:57.209474] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:39.752 [2024-12-05 18:58:57.209507] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:39.752 [2024-12-05 18:58:57.209518] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:39.752 [2024-12-05 18:58:57.209524] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:39.752 [2024-12-05 18:58:57.209535] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:39.752 "name": "Existed_Raid", 00:08:39.752 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.752 "strip_size_kb": 64, 00:08:39.752 "state": "configuring", 00:08:39.752 "raid_level": "concat", 00:08:39.752 "superblock": false, 00:08:39.752 "num_base_bdevs": 3, 00:08:39.752 "num_base_bdevs_discovered": 0, 00:08:39.752 "num_base_bdevs_operational": 3, 00:08:39.752 "base_bdevs_list": [ 00:08:39.752 { 00:08:39.752 "name": "BaseBdev1", 00:08:39.752 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.752 "is_configured": false, 00:08:39.752 "data_offset": 0, 00:08:39.752 "data_size": 0 00:08:39.752 }, 00:08:39.752 { 00:08:39.752 "name": "BaseBdev2", 00:08:39.752 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.752 "is_configured": false, 00:08:39.752 "data_offset": 0, 00:08:39.752 "data_size": 0 00:08:39.752 }, 00:08:39.752 { 00:08:39.752 "name": "BaseBdev3", 00:08:39.752 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.752 "is_configured": false, 00:08:39.752 "data_offset": 0, 00:08:39.752 "data_size": 0 00:08:39.752 } 00:08:39.752 ] 00:08:39.752 }' 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:39.752 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.319 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:40.319 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.320 [2024-12-05 18:58:57.676518] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:40.320 [2024-12-05 18:58:57.676603] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.320 [2024-12-05 18:58:57.684526] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:40.320 [2024-12-05 18:58:57.684623] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:40.320 [2024-12-05 18:58:57.684650] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:40.320 [2024-12-05 18:58:57.684689] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:40.320 [2024-12-05 18:58:57.684708] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:40.320 [2024-12-05 18:58:57.684728] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.320 [2024-12-05 18:58:57.701483] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:40.320 BaseBdev1 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.320 [ 00:08:40.320 { 00:08:40.320 "name": "BaseBdev1", 00:08:40.320 "aliases": [ 00:08:40.320 "681c76ff-e874-4a71-aced-bb13bf12fba5" 00:08:40.320 ], 00:08:40.320 "product_name": "Malloc disk", 00:08:40.320 "block_size": 512, 00:08:40.320 "num_blocks": 65536, 00:08:40.320 "uuid": "681c76ff-e874-4a71-aced-bb13bf12fba5", 00:08:40.320 "assigned_rate_limits": { 00:08:40.320 "rw_ios_per_sec": 0, 00:08:40.320 "rw_mbytes_per_sec": 0, 00:08:40.320 "r_mbytes_per_sec": 0, 00:08:40.320 "w_mbytes_per_sec": 0 00:08:40.320 }, 00:08:40.320 "claimed": true, 00:08:40.320 "claim_type": "exclusive_write", 00:08:40.320 "zoned": false, 00:08:40.320 "supported_io_types": { 00:08:40.320 "read": true, 00:08:40.320 "write": true, 00:08:40.320 "unmap": true, 00:08:40.320 "flush": true, 00:08:40.320 "reset": true, 00:08:40.320 "nvme_admin": false, 00:08:40.320 "nvme_io": false, 00:08:40.320 "nvme_io_md": false, 00:08:40.320 "write_zeroes": true, 00:08:40.320 "zcopy": true, 00:08:40.320 "get_zone_info": false, 00:08:40.320 "zone_management": false, 00:08:40.320 "zone_append": false, 00:08:40.320 "compare": false, 00:08:40.320 "compare_and_write": false, 00:08:40.320 "abort": true, 00:08:40.320 "seek_hole": false, 00:08:40.320 "seek_data": false, 00:08:40.320 "copy": true, 00:08:40.320 "nvme_iov_md": false 00:08:40.320 }, 00:08:40.320 "memory_domains": [ 00:08:40.320 { 00:08:40.320 "dma_device_id": "system", 00:08:40.320 "dma_device_type": 1 00:08:40.320 }, 00:08:40.320 { 00:08:40.320 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:40.320 "dma_device_type": 2 00:08:40.320 } 00:08:40.320 ], 00:08:40.320 "driver_specific": {} 00:08:40.320 } 00:08:40.320 ] 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:40.320 "name": "Existed_Raid", 00:08:40.320 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.320 "strip_size_kb": 64, 00:08:40.320 "state": "configuring", 00:08:40.320 "raid_level": "concat", 00:08:40.320 "superblock": false, 00:08:40.320 "num_base_bdevs": 3, 00:08:40.320 "num_base_bdevs_discovered": 1, 00:08:40.320 "num_base_bdevs_operational": 3, 00:08:40.320 "base_bdevs_list": [ 00:08:40.320 { 00:08:40.320 "name": "BaseBdev1", 00:08:40.320 "uuid": "681c76ff-e874-4a71-aced-bb13bf12fba5", 00:08:40.320 "is_configured": true, 00:08:40.320 "data_offset": 0, 00:08:40.320 "data_size": 65536 00:08:40.320 }, 00:08:40.320 { 00:08:40.320 "name": "BaseBdev2", 00:08:40.320 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.320 "is_configured": false, 00:08:40.320 "data_offset": 0, 00:08:40.320 "data_size": 0 00:08:40.320 }, 00:08:40.320 { 00:08:40.320 "name": "BaseBdev3", 00:08:40.320 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.320 "is_configured": false, 00:08:40.320 "data_offset": 0, 00:08:40.320 "data_size": 0 00:08:40.320 } 00:08:40.320 ] 00:08:40.320 }' 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:40.320 18:58:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.578 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:40.578 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.578 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.578 [2024-12-05 18:58:58.108825] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:40.578 [2024-12-05 18:58:58.108867] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.579 [2024-12-05 18:58:58.120845] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:40.579 [2024-12-05 18:58:58.122726] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:40.579 [2024-12-05 18:58:58.122811] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:40.579 [2024-12-05 18:58:58.122824] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:40.579 [2024-12-05 18:58:58.122836] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:40.579 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.837 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:40.837 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:40.837 "name": "Existed_Raid", 00:08:40.837 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.837 "strip_size_kb": 64, 00:08:40.837 "state": "configuring", 00:08:40.837 "raid_level": "concat", 00:08:40.837 "superblock": false, 00:08:40.837 "num_base_bdevs": 3, 00:08:40.837 "num_base_bdevs_discovered": 1, 00:08:40.837 "num_base_bdevs_operational": 3, 00:08:40.837 "base_bdevs_list": [ 00:08:40.837 { 00:08:40.837 "name": "BaseBdev1", 00:08:40.837 "uuid": "681c76ff-e874-4a71-aced-bb13bf12fba5", 00:08:40.837 "is_configured": true, 00:08:40.837 "data_offset": 0, 00:08:40.837 "data_size": 65536 00:08:40.837 }, 00:08:40.837 { 00:08:40.837 "name": "BaseBdev2", 00:08:40.837 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.837 "is_configured": false, 00:08:40.837 "data_offset": 0, 00:08:40.837 "data_size": 0 00:08:40.837 }, 00:08:40.837 { 00:08:40.837 "name": "BaseBdev3", 00:08:40.837 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:40.837 "is_configured": false, 00:08:40.837 "data_offset": 0, 00:08:40.837 "data_size": 0 00:08:40.837 } 00:08:40.837 ] 00:08:40.837 }' 00:08:40.837 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:40.837 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.095 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:41.095 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.095 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.095 [2024-12-05 18:58:58.559137] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:41.095 BaseBdev2 00:08:41.095 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.095 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:41.095 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:41.095 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:41.095 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:41.095 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:41.095 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:41.095 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:41.095 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.095 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.095 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.096 [ 00:08:41.096 { 00:08:41.096 "name": "BaseBdev2", 00:08:41.096 "aliases": [ 00:08:41.096 "86d9f772-3308-4578-9231-1e9b4a8a6b5e" 00:08:41.096 ], 00:08:41.096 "product_name": "Malloc disk", 00:08:41.096 "block_size": 512, 00:08:41.096 "num_blocks": 65536, 00:08:41.096 "uuid": "86d9f772-3308-4578-9231-1e9b4a8a6b5e", 00:08:41.096 "assigned_rate_limits": { 00:08:41.096 "rw_ios_per_sec": 0, 00:08:41.096 "rw_mbytes_per_sec": 0, 00:08:41.096 "r_mbytes_per_sec": 0, 00:08:41.096 "w_mbytes_per_sec": 0 00:08:41.096 }, 00:08:41.096 "claimed": true, 00:08:41.096 "claim_type": "exclusive_write", 00:08:41.096 "zoned": false, 00:08:41.096 "supported_io_types": { 00:08:41.096 "read": true, 00:08:41.096 "write": true, 00:08:41.096 "unmap": true, 00:08:41.096 "flush": true, 00:08:41.096 "reset": true, 00:08:41.096 "nvme_admin": false, 00:08:41.096 "nvme_io": false, 00:08:41.096 "nvme_io_md": false, 00:08:41.096 "write_zeroes": true, 00:08:41.096 "zcopy": true, 00:08:41.096 "get_zone_info": false, 00:08:41.096 "zone_management": false, 00:08:41.096 "zone_append": false, 00:08:41.096 "compare": false, 00:08:41.096 "compare_and_write": false, 00:08:41.096 "abort": true, 00:08:41.096 "seek_hole": false, 00:08:41.096 "seek_data": false, 00:08:41.096 "copy": true, 00:08:41.096 "nvme_iov_md": false 00:08:41.096 }, 00:08:41.096 "memory_domains": [ 00:08:41.096 { 00:08:41.096 "dma_device_id": "system", 00:08:41.096 "dma_device_type": 1 00:08:41.096 }, 00:08:41.096 { 00:08:41.096 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:41.096 "dma_device_type": 2 00:08:41.096 } 00:08:41.096 ], 00:08:41.096 "driver_specific": {} 00:08:41.096 } 00:08:41.096 ] 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:41.096 "name": "Existed_Raid", 00:08:41.096 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:41.096 "strip_size_kb": 64, 00:08:41.096 "state": "configuring", 00:08:41.096 "raid_level": "concat", 00:08:41.096 "superblock": false, 00:08:41.096 "num_base_bdevs": 3, 00:08:41.096 "num_base_bdevs_discovered": 2, 00:08:41.096 "num_base_bdevs_operational": 3, 00:08:41.096 "base_bdevs_list": [ 00:08:41.096 { 00:08:41.096 "name": "BaseBdev1", 00:08:41.096 "uuid": "681c76ff-e874-4a71-aced-bb13bf12fba5", 00:08:41.096 "is_configured": true, 00:08:41.096 "data_offset": 0, 00:08:41.096 "data_size": 65536 00:08:41.096 }, 00:08:41.096 { 00:08:41.096 "name": "BaseBdev2", 00:08:41.096 "uuid": "86d9f772-3308-4578-9231-1e9b4a8a6b5e", 00:08:41.096 "is_configured": true, 00:08:41.096 "data_offset": 0, 00:08:41.096 "data_size": 65536 00:08:41.096 }, 00:08:41.096 { 00:08:41.096 "name": "BaseBdev3", 00:08:41.096 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:41.096 "is_configured": false, 00:08:41.096 "data_offset": 0, 00:08:41.096 "data_size": 0 00:08:41.096 } 00:08:41.096 ] 00:08:41.096 }' 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:41.096 18:58:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.664 [2024-12-05 18:58:59.059918] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:41.664 [2024-12-05 18:58:59.060014] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:41.664 [2024-12-05 18:58:59.060047] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:41.664 [2024-12-05 18:58:59.061006] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:41.664 [2024-12-05 18:58:59.061446] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:41.664 [2024-12-05 18:58:59.061498] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:41.664 BaseBdev3 00:08:41.664 [2024-12-05 18:58:59.062107] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.664 [ 00:08:41.664 { 00:08:41.664 "name": "BaseBdev3", 00:08:41.664 "aliases": [ 00:08:41.664 "72b9158f-7301-4968-b789-04d9b7eab6e7" 00:08:41.664 ], 00:08:41.664 "product_name": "Malloc disk", 00:08:41.664 "block_size": 512, 00:08:41.664 "num_blocks": 65536, 00:08:41.664 "uuid": "72b9158f-7301-4968-b789-04d9b7eab6e7", 00:08:41.664 "assigned_rate_limits": { 00:08:41.664 "rw_ios_per_sec": 0, 00:08:41.664 "rw_mbytes_per_sec": 0, 00:08:41.664 "r_mbytes_per_sec": 0, 00:08:41.664 "w_mbytes_per_sec": 0 00:08:41.664 }, 00:08:41.664 "claimed": true, 00:08:41.664 "claim_type": "exclusive_write", 00:08:41.664 "zoned": false, 00:08:41.664 "supported_io_types": { 00:08:41.664 "read": true, 00:08:41.664 "write": true, 00:08:41.664 "unmap": true, 00:08:41.664 "flush": true, 00:08:41.664 "reset": true, 00:08:41.664 "nvme_admin": false, 00:08:41.664 "nvme_io": false, 00:08:41.664 "nvme_io_md": false, 00:08:41.664 "write_zeroes": true, 00:08:41.664 "zcopy": true, 00:08:41.664 "get_zone_info": false, 00:08:41.664 "zone_management": false, 00:08:41.664 "zone_append": false, 00:08:41.664 "compare": false, 00:08:41.664 "compare_and_write": false, 00:08:41.664 "abort": true, 00:08:41.664 "seek_hole": false, 00:08:41.664 "seek_data": false, 00:08:41.664 "copy": true, 00:08:41.664 "nvme_iov_md": false 00:08:41.664 }, 00:08:41.664 "memory_domains": [ 00:08:41.664 { 00:08:41.664 "dma_device_id": "system", 00:08:41.664 "dma_device_type": 1 00:08:41.664 }, 00:08:41.664 { 00:08:41.664 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:41.664 "dma_device_type": 2 00:08:41.664 } 00:08:41.664 ], 00:08:41.664 "driver_specific": {} 00:08:41.664 } 00:08:41.664 ] 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:41.664 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:41.664 "name": "Existed_Raid", 00:08:41.664 "uuid": "1dea0a5f-0736-43f6-aab3-359b6cffc7b1", 00:08:41.664 "strip_size_kb": 64, 00:08:41.664 "state": "online", 00:08:41.664 "raid_level": "concat", 00:08:41.664 "superblock": false, 00:08:41.664 "num_base_bdevs": 3, 00:08:41.664 "num_base_bdevs_discovered": 3, 00:08:41.664 "num_base_bdevs_operational": 3, 00:08:41.664 "base_bdevs_list": [ 00:08:41.664 { 00:08:41.664 "name": "BaseBdev1", 00:08:41.664 "uuid": "681c76ff-e874-4a71-aced-bb13bf12fba5", 00:08:41.665 "is_configured": true, 00:08:41.665 "data_offset": 0, 00:08:41.665 "data_size": 65536 00:08:41.665 }, 00:08:41.665 { 00:08:41.665 "name": "BaseBdev2", 00:08:41.665 "uuid": "86d9f772-3308-4578-9231-1e9b4a8a6b5e", 00:08:41.665 "is_configured": true, 00:08:41.665 "data_offset": 0, 00:08:41.665 "data_size": 65536 00:08:41.665 }, 00:08:41.665 { 00:08:41.665 "name": "BaseBdev3", 00:08:41.665 "uuid": "72b9158f-7301-4968-b789-04d9b7eab6e7", 00:08:41.665 "is_configured": true, 00:08:41.665 "data_offset": 0, 00:08:41.665 "data_size": 65536 00:08:41.665 } 00:08:41.665 ] 00:08:41.665 }' 00:08:41.665 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:41.665 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.924 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:41.924 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:41.924 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:41.924 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:41.924 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:41.924 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:41.924 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:41.924 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:41.924 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:41.924 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.924 [2024-12-05 18:58:59.479449] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:42.185 "name": "Existed_Raid", 00:08:42.185 "aliases": [ 00:08:42.185 "1dea0a5f-0736-43f6-aab3-359b6cffc7b1" 00:08:42.185 ], 00:08:42.185 "product_name": "Raid Volume", 00:08:42.185 "block_size": 512, 00:08:42.185 "num_blocks": 196608, 00:08:42.185 "uuid": "1dea0a5f-0736-43f6-aab3-359b6cffc7b1", 00:08:42.185 "assigned_rate_limits": { 00:08:42.185 "rw_ios_per_sec": 0, 00:08:42.185 "rw_mbytes_per_sec": 0, 00:08:42.185 "r_mbytes_per_sec": 0, 00:08:42.185 "w_mbytes_per_sec": 0 00:08:42.185 }, 00:08:42.185 "claimed": false, 00:08:42.185 "zoned": false, 00:08:42.185 "supported_io_types": { 00:08:42.185 "read": true, 00:08:42.185 "write": true, 00:08:42.185 "unmap": true, 00:08:42.185 "flush": true, 00:08:42.185 "reset": true, 00:08:42.185 "nvme_admin": false, 00:08:42.185 "nvme_io": false, 00:08:42.185 "nvme_io_md": false, 00:08:42.185 "write_zeroes": true, 00:08:42.185 "zcopy": false, 00:08:42.185 "get_zone_info": false, 00:08:42.185 "zone_management": false, 00:08:42.185 "zone_append": false, 00:08:42.185 "compare": false, 00:08:42.185 "compare_and_write": false, 00:08:42.185 "abort": false, 00:08:42.185 "seek_hole": false, 00:08:42.185 "seek_data": false, 00:08:42.185 "copy": false, 00:08:42.185 "nvme_iov_md": false 00:08:42.185 }, 00:08:42.185 "memory_domains": [ 00:08:42.185 { 00:08:42.185 "dma_device_id": "system", 00:08:42.185 "dma_device_type": 1 00:08:42.185 }, 00:08:42.185 { 00:08:42.185 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.185 "dma_device_type": 2 00:08:42.185 }, 00:08:42.185 { 00:08:42.185 "dma_device_id": "system", 00:08:42.185 "dma_device_type": 1 00:08:42.185 }, 00:08:42.185 { 00:08:42.185 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.185 "dma_device_type": 2 00:08:42.185 }, 00:08:42.185 { 00:08:42.185 "dma_device_id": "system", 00:08:42.185 "dma_device_type": 1 00:08:42.185 }, 00:08:42.185 { 00:08:42.185 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.185 "dma_device_type": 2 00:08:42.185 } 00:08:42.185 ], 00:08:42.185 "driver_specific": { 00:08:42.185 "raid": { 00:08:42.185 "uuid": "1dea0a5f-0736-43f6-aab3-359b6cffc7b1", 00:08:42.185 "strip_size_kb": 64, 00:08:42.185 "state": "online", 00:08:42.185 "raid_level": "concat", 00:08:42.185 "superblock": false, 00:08:42.185 "num_base_bdevs": 3, 00:08:42.185 "num_base_bdevs_discovered": 3, 00:08:42.185 "num_base_bdevs_operational": 3, 00:08:42.185 "base_bdevs_list": [ 00:08:42.185 { 00:08:42.185 "name": "BaseBdev1", 00:08:42.185 "uuid": "681c76ff-e874-4a71-aced-bb13bf12fba5", 00:08:42.185 "is_configured": true, 00:08:42.185 "data_offset": 0, 00:08:42.185 "data_size": 65536 00:08:42.185 }, 00:08:42.185 { 00:08:42.185 "name": "BaseBdev2", 00:08:42.185 "uuid": "86d9f772-3308-4578-9231-1e9b4a8a6b5e", 00:08:42.185 "is_configured": true, 00:08:42.185 "data_offset": 0, 00:08:42.185 "data_size": 65536 00:08:42.185 }, 00:08:42.185 { 00:08:42.185 "name": "BaseBdev3", 00:08:42.185 "uuid": "72b9158f-7301-4968-b789-04d9b7eab6e7", 00:08:42.185 "is_configured": true, 00:08:42.185 "data_offset": 0, 00:08:42.185 "data_size": 65536 00:08:42.185 } 00:08:42.185 ] 00:08:42.185 } 00:08:42.185 } 00:08:42.185 }' 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:42.185 BaseBdev2 00:08:42.185 BaseBdev3' 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.185 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.446 [2024-12-05 18:58:59.762734] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:42.446 [2024-12-05 18:58:59.762801] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:42.446 [2024-12-05 18:58:59.762890] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:42.446 "name": "Existed_Raid", 00:08:42.446 "uuid": "1dea0a5f-0736-43f6-aab3-359b6cffc7b1", 00:08:42.446 "strip_size_kb": 64, 00:08:42.446 "state": "offline", 00:08:42.446 "raid_level": "concat", 00:08:42.446 "superblock": false, 00:08:42.446 "num_base_bdevs": 3, 00:08:42.446 "num_base_bdevs_discovered": 2, 00:08:42.446 "num_base_bdevs_operational": 2, 00:08:42.446 "base_bdevs_list": [ 00:08:42.446 { 00:08:42.446 "name": null, 00:08:42.446 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:42.446 "is_configured": false, 00:08:42.446 "data_offset": 0, 00:08:42.446 "data_size": 65536 00:08:42.446 }, 00:08:42.446 { 00:08:42.446 "name": "BaseBdev2", 00:08:42.446 "uuid": "86d9f772-3308-4578-9231-1e9b4a8a6b5e", 00:08:42.446 "is_configured": true, 00:08:42.446 "data_offset": 0, 00:08:42.446 "data_size": 65536 00:08:42.446 }, 00:08:42.446 { 00:08:42.446 "name": "BaseBdev3", 00:08:42.446 "uuid": "72b9158f-7301-4968-b789-04d9b7eab6e7", 00:08:42.446 "is_configured": true, 00:08:42.446 "data_offset": 0, 00:08:42.446 "data_size": 65536 00:08:42.446 } 00:08:42.446 ] 00:08:42.446 }' 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:42.446 18:58:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.706 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:42.706 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:42.706 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:42.706 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.706 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.706 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.706 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.706 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:42.706 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:42.706 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:42.706 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.706 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.706 [2024-12-05 18:59:00.253180] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.967 [2024-12-05 18:59:00.320208] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:42.967 [2024-12-05 18:59:00.320255] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.967 BaseBdev2 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.967 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.967 [ 00:08:42.967 { 00:08:42.967 "name": "BaseBdev2", 00:08:42.967 "aliases": [ 00:08:42.967 "63fbc760-dc3f-47bc-b066-8724b64a27ed" 00:08:42.967 ], 00:08:42.967 "product_name": "Malloc disk", 00:08:42.967 "block_size": 512, 00:08:42.967 "num_blocks": 65536, 00:08:42.967 "uuid": "63fbc760-dc3f-47bc-b066-8724b64a27ed", 00:08:42.967 "assigned_rate_limits": { 00:08:42.967 "rw_ios_per_sec": 0, 00:08:42.967 "rw_mbytes_per_sec": 0, 00:08:42.968 "r_mbytes_per_sec": 0, 00:08:42.968 "w_mbytes_per_sec": 0 00:08:42.968 }, 00:08:42.968 "claimed": false, 00:08:42.968 "zoned": false, 00:08:42.968 "supported_io_types": { 00:08:42.968 "read": true, 00:08:42.968 "write": true, 00:08:42.968 "unmap": true, 00:08:42.968 "flush": true, 00:08:42.968 "reset": true, 00:08:42.968 "nvme_admin": false, 00:08:42.968 "nvme_io": false, 00:08:42.968 "nvme_io_md": false, 00:08:42.968 "write_zeroes": true, 00:08:42.968 "zcopy": true, 00:08:42.968 "get_zone_info": false, 00:08:42.968 "zone_management": false, 00:08:42.968 "zone_append": false, 00:08:42.968 "compare": false, 00:08:42.968 "compare_and_write": false, 00:08:42.968 "abort": true, 00:08:42.968 "seek_hole": false, 00:08:42.968 "seek_data": false, 00:08:42.968 "copy": true, 00:08:42.968 "nvme_iov_md": false 00:08:42.968 }, 00:08:42.968 "memory_domains": [ 00:08:42.968 { 00:08:42.968 "dma_device_id": "system", 00:08:42.968 "dma_device_type": 1 00:08:42.968 }, 00:08:42.968 { 00:08:42.968 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.968 "dma_device_type": 2 00:08:42.968 } 00:08:42.968 ], 00:08:42.968 "driver_specific": {} 00:08:42.968 } 00:08:42.968 ] 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.968 BaseBdev3 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.968 [ 00:08:42.968 { 00:08:42.968 "name": "BaseBdev3", 00:08:42.968 "aliases": [ 00:08:42.968 "817398c4-fd4e-43b0-bbd3-baec7642d64a" 00:08:42.968 ], 00:08:42.968 "product_name": "Malloc disk", 00:08:42.968 "block_size": 512, 00:08:42.968 "num_blocks": 65536, 00:08:42.968 "uuid": "817398c4-fd4e-43b0-bbd3-baec7642d64a", 00:08:42.968 "assigned_rate_limits": { 00:08:42.968 "rw_ios_per_sec": 0, 00:08:42.968 "rw_mbytes_per_sec": 0, 00:08:42.968 "r_mbytes_per_sec": 0, 00:08:42.968 "w_mbytes_per_sec": 0 00:08:42.968 }, 00:08:42.968 "claimed": false, 00:08:42.968 "zoned": false, 00:08:42.968 "supported_io_types": { 00:08:42.968 "read": true, 00:08:42.968 "write": true, 00:08:42.968 "unmap": true, 00:08:42.968 "flush": true, 00:08:42.968 "reset": true, 00:08:42.968 "nvme_admin": false, 00:08:42.968 "nvme_io": false, 00:08:42.968 "nvme_io_md": false, 00:08:42.968 "write_zeroes": true, 00:08:42.968 "zcopy": true, 00:08:42.968 "get_zone_info": false, 00:08:42.968 "zone_management": false, 00:08:42.968 "zone_append": false, 00:08:42.968 "compare": false, 00:08:42.968 "compare_and_write": false, 00:08:42.968 "abort": true, 00:08:42.968 "seek_hole": false, 00:08:42.968 "seek_data": false, 00:08:42.968 "copy": true, 00:08:42.968 "nvme_iov_md": false 00:08:42.968 }, 00:08:42.968 "memory_domains": [ 00:08:42.968 { 00:08:42.968 "dma_device_id": "system", 00:08:42.968 "dma_device_type": 1 00:08:42.968 }, 00:08:42.968 { 00:08:42.968 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.968 "dma_device_type": 2 00:08:42.968 } 00:08:42.968 ], 00:08:42.968 "driver_specific": {} 00:08:42.968 } 00:08:42.968 ] 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.968 [2024-12-05 18:59:00.495137] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:42.968 [2024-12-05 18:59:00.495239] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:42.968 [2024-12-05 18:59:00.495281] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:42.968 [2024-12-05 18:59:00.497113] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:42.968 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:42.969 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:42.969 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:42.969 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:42.969 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:42.969 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:42.969 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.969 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:42.969 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.969 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:42.969 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.227 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:43.227 "name": "Existed_Raid", 00:08:43.227 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:43.227 "strip_size_kb": 64, 00:08:43.227 "state": "configuring", 00:08:43.227 "raid_level": "concat", 00:08:43.227 "superblock": false, 00:08:43.227 "num_base_bdevs": 3, 00:08:43.227 "num_base_bdevs_discovered": 2, 00:08:43.227 "num_base_bdevs_operational": 3, 00:08:43.227 "base_bdevs_list": [ 00:08:43.227 { 00:08:43.227 "name": "BaseBdev1", 00:08:43.227 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:43.227 "is_configured": false, 00:08:43.227 "data_offset": 0, 00:08:43.227 "data_size": 0 00:08:43.227 }, 00:08:43.227 { 00:08:43.227 "name": "BaseBdev2", 00:08:43.227 "uuid": "63fbc760-dc3f-47bc-b066-8724b64a27ed", 00:08:43.227 "is_configured": true, 00:08:43.227 "data_offset": 0, 00:08:43.227 "data_size": 65536 00:08:43.227 }, 00:08:43.227 { 00:08:43.227 "name": "BaseBdev3", 00:08:43.227 "uuid": "817398c4-fd4e-43b0-bbd3-baec7642d64a", 00:08:43.227 "is_configured": true, 00:08:43.227 "data_offset": 0, 00:08:43.227 "data_size": 65536 00:08:43.227 } 00:08:43.227 ] 00:08:43.227 }' 00:08:43.227 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:43.227 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.487 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:43.487 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.487 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.487 [2024-12-05 18:59:00.890459] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:43.487 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.487 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:43.487 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:43.488 "name": "Existed_Raid", 00:08:43.488 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:43.488 "strip_size_kb": 64, 00:08:43.488 "state": "configuring", 00:08:43.488 "raid_level": "concat", 00:08:43.488 "superblock": false, 00:08:43.488 "num_base_bdevs": 3, 00:08:43.488 "num_base_bdevs_discovered": 1, 00:08:43.488 "num_base_bdevs_operational": 3, 00:08:43.488 "base_bdevs_list": [ 00:08:43.488 { 00:08:43.488 "name": "BaseBdev1", 00:08:43.488 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:43.488 "is_configured": false, 00:08:43.488 "data_offset": 0, 00:08:43.488 "data_size": 0 00:08:43.488 }, 00:08:43.488 { 00:08:43.488 "name": null, 00:08:43.488 "uuid": "63fbc760-dc3f-47bc-b066-8724b64a27ed", 00:08:43.488 "is_configured": false, 00:08:43.488 "data_offset": 0, 00:08:43.488 "data_size": 65536 00:08:43.488 }, 00:08:43.488 { 00:08:43.488 "name": "BaseBdev3", 00:08:43.488 "uuid": "817398c4-fd4e-43b0-bbd3-baec7642d64a", 00:08:43.488 "is_configured": true, 00:08:43.488 "data_offset": 0, 00:08:43.488 "data_size": 65536 00:08:43.488 } 00:08:43.488 ] 00:08:43.488 }' 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:43.488 18:59:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.748 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:43.748 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:43.748 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:43.748 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.748 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.008 [2024-12-05 18:59:01.332799] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:44.008 BaseBdev1 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.008 [ 00:08:44.008 { 00:08:44.008 "name": "BaseBdev1", 00:08:44.008 "aliases": [ 00:08:44.008 "395db821-14cf-40e8-a6f5-47ef4f3f805d" 00:08:44.008 ], 00:08:44.008 "product_name": "Malloc disk", 00:08:44.008 "block_size": 512, 00:08:44.008 "num_blocks": 65536, 00:08:44.008 "uuid": "395db821-14cf-40e8-a6f5-47ef4f3f805d", 00:08:44.008 "assigned_rate_limits": { 00:08:44.008 "rw_ios_per_sec": 0, 00:08:44.008 "rw_mbytes_per_sec": 0, 00:08:44.008 "r_mbytes_per_sec": 0, 00:08:44.008 "w_mbytes_per_sec": 0 00:08:44.008 }, 00:08:44.008 "claimed": true, 00:08:44.008 "claim_type": "exclusive_write", 00:08:44.008 "zoned": false, 00:08:44.008 "supported_io_types": { 00:08:44.008 "read": true, 00:08:44.008 "write": true, 00:08:44.008 "unmap": true, 00:08:44.008 "flush": true, 00:08:44.008 "reset": true, 00:08:44.008 "nvme_admin": false, 00:08:44.008 "nvme_io": false, 00:08:44.008 "nvme_io_md": false, 00:08:44.008 "write_zeroes": true, 00:08:44.008 "zcopy": true, 00:08:44.008 "get_zone_info": false, 00:08:44.008 "zone_management": false, 00:08:44.008 "zone_append": false, 00:08:44.008 "compare": false, 00:08:44.008 "compare_and_write": false, 00:08:44.008 "abort": true, 00:08:44.008 "seek_hole": false, 00:08:44.008 "seek_data": false, 00:08:44.008 "copy": true, 00:08:44.008 "nvme_iov_md": false 00:08:44.008 }, 00:08:44.008 "memory_domains": [ 00:08:44.008 { 00:08:44.008 "dma_device_id": "system", 00:08:44.008 "dma_device_type": 1 00:08:44.008 }, 00:08:44.008 { 00:08:44.008 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:44.008 "dma_device_type": 2 00:08:44.008 } 00:08:44.008 ], 00:08:44.008 "driver_specific": {} 00:08:44.008 } 00:08:44.008 ] 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.008 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:44.008 "name": "Existed_Raid", 00:08:44.008 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.008 "strip_size_kb": 64, 00:08:44.009 "state": "configuring", 00:08:44.009 "raid_level": "concat", 00:08:44.009 "superblock": false, 00:08:44.009 "num_base_bdevs": 3, 00:08:44.009 "num_base_bdevs_discovered": 2, 00:08:44.009 "num_base_bdevs_operational": 3, 00:08:44.009 "base_bdevs_list": [ 00:08:44.009 { 00:08:44.009 "name": "BaseBdev1", 00:08:44.009 "uuid": "395db821-14cf-40e8-a6f5-47ef4f3f805d", 00:08:44.009 "is_configured": true, 00:08:44.009 "data_offset": 0, 00:08:44.009 "data_size": 65536 00:08:44.009 }, 00:08:44.009 { 00:08:44.009 "name": null, 00:08:44.009 "uuid": "63fbc760-dc3f-47bc-b066-8724b64a27ed", 00:08:44.009 "is_configured": false, 00:08:44.009 "data_offset": 0, 00:08:44.009 "data_size": 65536 00:08:44.009 }, 00:08:44.009 { 00:08:44.009 "name": "BaseBdev3", 00:08:44.009 "uuid": "817398c4-fd4e-43b0-bbd3-baec7642d64a", 00:08:44.009 "is_configured": true, 00:08:44.009 "data_offset": 0, 00:08:44.009 "data_size": 65536 00:08:44.009 } 00:08:44.009 ] 00:08:44.009 }' 00:08:44.009 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:44.009 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.269 [2024-12-05 18:59:01.812061] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.269 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.529 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.529 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:44.529 "name": "Existed_Raid", 00:08:44.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.529 "strip_size_kb": 64, 00:08:44.529 "state": "configuring", 00:08:44.529 "raid_level": "concat", 00:08:44.529 "superblock": false, 00:08:44.529 "num_base_bdevs": 3, 00:08:44.529 "num_base_bdevs_discovered": 1, 00:08:44.529 "num_base_bdevs_operational": 3, 00:08:44.529 "base_bdevs_list": [ 00:08:44.529 { 00:08:44.529 "name": "BaseBdev1", 00:08:44.529 "uuid": "395db821-14cf-40e8-a6f5-47ef4f3f805d", 00:08:44.529 "is_configured": true, 00:08:44.529 "data_offset": 0, 00:08:44.529 "data_size": 65536 00:08:44.529 }, 00:08:44.529 { 00:08:44.529 "name": null, 00:08:44.529 "uuid": "63fbc760-dc3f-47bc-b066-8724b64a27ed", 00:08:44.529 "is_configured": false, 00:08:44.529 "data_offset": 0, 00:08:44.529 "data_size": 65536 00:08:44.529 }, 00:08:44.529 { 00:08:44.529 "name": null, 00:08:44.529 "uuid": "817398c4-fd4e-43b0-bbd3-baec7642d64a", 00:08:44.529 "is_configured": false, 00:08:44.529 "data_offset": 0, 00:08:44.529 "data_size": 65536 00:08:44.529 } 00:08:44.529 ] 00:08:44.529 }' 00:08:44.529 18:59:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:44.529 18:59:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.789 [2024-12-05 18:59:02.267293] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:44.789 "name": "Existed_Raid", 00:08:44.789 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:44.789 "strip_size_kb": 64, 00:08:44.789 "state": "configuring", 00:08:44.789 "raid_level": "concat", 00:08:44.789 "superblock": false, 00:08:44.789 "num_base_bdevs": 3, 00:08:44.789 "num_base_bdevs_discovered": 2, 00:08:44.789 "num_base_bdevs_operational": 3, 00:08:44.789 "base_bdevs_list": [ 00:08:44.789 { 00:08:44.789 "name": "BaseBdev1", 00:08:44.789 "uuid": "395db821-14cf-40e8-a6f5-47ef4f3f805d", 00:08:44.789 "is_configured": true, 00:08:44.789 "data_offset": 0, 00:08:44.789 "data_size": 65536 00:08:44.789 }, 00:08:44.789 { 00:08:44.789 "name": null, 00:08:44.789 "uuid": "63fbc760-dc3f-47bc-b066-8724b64a27ed", 00:08:44.789 "is_configured": false, 00:08:44.789 "data_offset": 0, 00:08:44.789 "data_size": 65536 00:08:44.789 }, 00:08:44.789 { 00:08:44.789 "name": "BaseBdev3", 00:08:44.789 "uuid": "817398c4-fd4e-43b0-bbd3-baec7642d64a", 00:08:44.789 "is_configured": true, 00:08:44.789 "data_offset": 0, 00:08:44.789 "data_size": 65536 00:08:44.789 } 00:08:44.789 ] 00:08:44.789 }' 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:44.789 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.358 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.358 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.358 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.358 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:45.358 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.358 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:45.358 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:45.358 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.358 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.358 [2024-12-05 18:59:02.750488] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:45.358 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.358 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:45.358 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:45.358 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:45.358 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:45.359 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:45.359 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:45.359 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:45.359 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:45.359 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:45.359 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:45.359 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.359 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:45.359 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.359 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.359 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.359 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:45.359 "name": "Existed_Raid", 00:08:45.359 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:45.359 "strip_size_kb": 64, 00:08:45.359 "state": "configuring", 00:08:45.359 "raid_level": "concat", 00:08:45.359 "superblock": false, 00:08:45.359 "num_base_bdevs": 3, 00:08:45.359 "num_base_bdevs_discovered": 1, 00:08:45.359 "num_base_bdevs_operational": 3, 00:08:45.359 "base_bdevs_list": [ 00:08:45.359 { 00:08:45.359 "name": null, 00:08:45.359 "uuid": "395db821-14cf-40e8-a6f5-47ef4f3f805d", 00:08:45.359 "is_configured": false, 00:08:45.359 "data_offset": 0, 00:08:45.359 "data_size": 65536 00:08:45.359 }, 00:08:45.359 { 00:08:45.359 "name": null, 00:08:45.359 "uuid": "63fbc760-dc3f-47bc-b066-8724b64a27ed", 00:08:45.359 "is_configured": false, 00:08:45.359 "data_offset": 0, 00:08:45.359 "data_size": 65536 00:08:45.359 }, 00:08:45.359 { 00:08:45.359 "name": "BaseBdev3", 00:08:45.359 "uuid": "817398c4-fd4e-43b0-bbd3-baec7642d64a", 00:08:45.359 "is_configured": true, 00:08:45.359 "data_offset": 0, 00:08:45.359 "data_size": 65536 00:08:45.359 } 00:08:45.359 ] 00:08:45.359 }' 00:08:45.359 18:59:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:45.359 18:59:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.929 [2024-12-05 18:59:03.259988] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:45.929 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:45.929 "name": "Existed_Raid", 00:08:45.929 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:45.929 "strip_size_kb": 64, 00:08:45.929 "state": "configuring", 00:08:45.929 "raid_level": "concat", 00:08:45.929 "superblock": false, 00:08:45.929 "num_base_bdevs": 3, 00:08:45.929 "num_base_bdevs_discovered": 2, 00:08:45.929 "num_base_bdevs_operational": 3, 00:08:45.929 "base_bdevs_list": [ 00:08:45.929 { 00:08:45.929 "name": null, 00:08:45.929 "uuid": "395db821-14cf-40e8-a6f5-47ef4f3f805d", 00:08:45.929 "is_configured": false, 00:08:45.929 "data_offset": 0, 00:08:45.929 "data_size": 65536 00:08:45.929 }, 00:08:45.929 { 00:08:45.929 "name": "BaseBdev2", 00:08:45.929 "uuid": "63fbc760-dc3f-47bc-b066-8724b64a27ed", 00:08:45.929 "is_configured": true, 00:08:45.929 "data_offset": 0, 00:08:45.929 "data_size": 65536 00:08:45.929 }, 00:08:45.929 { 00:08:45.929 "name": "BaseBdev3", 00:08:45.930 "uuid": "817398c4-fd4e-43b0-bbd3-baec7642d64a", 00:08:45.930 "is_configured": true, 00:08:45.930 "data_offset": 0, 00:08:45.930 "data_size": 65536 00:08:45.930 } 00:08:45.930 ] 00:08:45.930 }' 00:08:45.930 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:45.930 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.188 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:46.188 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:46.188 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.188 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.188 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.188 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:46.188 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:46.188 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.188 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.188 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 395db821-14cf-40e8-a6f5-47ef4f3f805d 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.447 [2024-12-05 18:59:03.798062] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:46.447 [2024-12-05 18:59:03.798105] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:46.447 [2024-12-05 18:59:03.798114] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:46.447 [2024-12-05 18:59:03.798361] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:46.447 [2024-12-05 18:59:03.798481] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:46.447 [2024-12-05 18:59:03.798489] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:46.447 [2024-12-05 18:59:03.798697] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:46.447 NewBaseBdev 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.447 [ 00:08:46.447 { 00:08:46.447 "name": "NewBaseBdev", 00:08:46.447 "aliases": [ 00:08:46.447 "395db821-14cf-40e8-a6f5-47ef4f3f805d" 00:08:46.447 ], 00:08:46.447 "product_name": "Malloc disk", 00:08:46.447 "block_size": 512, 00:08:46.447 "num_blocks": 65536, 00:08:46.447 "uuid": "395db821-14cf-40e8-a6f5-47ef4f3f805d", 00:08:46.447 "assigned_rate_limits": { 00:08:46.447 "rw_ios_per_sec": 0, 00:08:46.447 "rw_mbytes_per_sec": 0, 00:08:46.447 "r_mbytes_per_sec": 0, 00:08:46.447 "w_mbytes_per_sec": 0 00:08:46.447 }, 00:08:46.447 "claimed": true, 00:08:46.447 "claim_type": "exclusive_write", 00:08:46.447 "zoned": false, 00:08:46.447 "supported_io_types": { 00:08:46.447 "read": true, 00:08:46.447 "write": true, 00:08:46.447 "unmap": true, 00:08:46.447 "flush": true, 00:08:46.447 "reset": true, 00:08:46.447 "nvme_admin": false, 00:08:46.447 "nvme_io": false, 00:08:46.447 "nvme_io_md": false, 00:08:46.447 "write_zeroes": true, 00:08:46.447 "zcopy": true, 00:08:46.447 "get_zone_info": false, 00:08:46.447 "zone_management": false, 00:08:46.447 "zone_append": false, 00:08:46.447 "compare": false, 00:08:46.447 "compare_and_write": false, 00:08:46.447 "abort": true, 00:08:46.447 "seek_hole": false, 00:08:46.447 "seek_data": false, 00:08:46.447 "copy": true, 00:08:46.447 "nvme_iov_md": false 00:08:46.447 }, 00:08:46.447 "memory_domains": [ 00:08:46.447 { 00:08:46.447 "dma_device_id": "system", 00:08:46.447 "dma_device_type": 1 00:08:46.447 }, 00:08:46.447 { 00:08:46.447 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.447 "dma_device_type": 2 00:08:46.447 } 00:08:46.447 ], 00:08:46.447 "driver_specific": {} 00:08:46.447 } 00:08:46.447 ] 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:46.447 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:46.448 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:46.448 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:46.448 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:46.448 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:46.448 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:46.448 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:46.448 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.448 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.448 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:46.448 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.448 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:46.448 "name": "Existed_Raid", 00:08:46.448 "uuid": "b15423b5-156f-4a5c-8d6d-d43766782e25", 00:08:46.448 "strip_size_kb": 64, 00:08:46.448 "state": "online", 00:08:46.448 "raid_level": "concat", 00:08:46.448 "superblock": false, 00:08:46.448 "num_base_bdevs": 3, 00:08:46.448 "num_base_bdevs_discovered": 3, 00:08:46.448 "num_base_bdevs_operational": 3, 00:08:46.448 "base_bdevs_list": [ 00:08:46.448 { 00:08:46.448 "name": "NewBaseBdev", 00:08:46.448 "uuid": "395db821-14cf-40e8-a6f5-47ef4f3f805d", 00:08:46.448 "is_configured": true, 00:08:46.448 "data_offset": 0, 00:08:46.448 "data_size": 65536 00:08:46.448 }, 00:08:46.448 { 00:08:46.448 "name": "BaseBdev2", 00:08:46.448 "uuid": "63fbc760-dc3f-47bc-b066-8724b64a27ed", 00:08:46.448 "is_configured": true, 00:08:46.448 "data_offset": 0, 00:08:46.448 "data_size": 65536 00:08:46.448 }, 00:08:46.448 { 00:08:46.448 "name": "BaseBdev3", 00:08:46.448 "uuid": "817398c4-fd4e-43b0-bbd3-baec7642d64a", 00:08:46.448 "is_configured": true, 00:08:46.448 "data_offset": 0, 00:08:46.448 "data_size": 65536 00:08:46.448 } 00:08:46.448 ] 00:08:46.448 }' 00:08:46.448 18:59:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:46.448 18:59:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.707 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:46.707 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:46.707 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:46.707 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:46.707 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:46.707 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:46.707 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:46.707 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:46.707 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.707 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.707 [2024-12-05 18:59:04.265538] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:46.967 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.967 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:46.967 "name": "Existed_Raid", 00:08:46.967 "aliases": [ 00:08:46.967 "b15423b5-156f-4a5c-8d6d-d43766782e25" 00:08:46.967 ], 00:08:46.967 "product_name": "Raid Volume", 00:08:46.967 "block_size": 512, 00:08:46.967 "num_blocks": 196608, 00:08:46.967 "uuid": "b15423b5-156f-4a5c-8d6d-d43766782e25", 00:08:46.967 "assigned_rate_limits": { 00:08:46.967 "rw_ios_per_sec": 0, 00:08:46.967 "rw_mbytes_per_sec": 0, 00:08:46.967 "r_mbytes_per_sec": 0, 00:08:46.967 "w_mbytes_per_sec": 0 00:08:46.967 }, 00:08:46.967 "claimed": false, 00:08:46.967 "zoned": false, 00:08:46.967 "supported_io_types": { 00:08:46.967 "read": true, 00:08:46.967 "write": true, 00:08:46.967 "unmap": true, 00:08:46.967 "flush": true, 00:08:46.967 "reset": true, 00:08:46.967 "nvme_admin": false, 00:08:46.967 "nvme_io": false, 00:08:46.967 "nvme_io_md": false, 00:08:46.967 "write_zeroes": true, 00:08:46.967 "zcopy": false, 00:08:46.967 "get_zone_info": false, 00:08:46.967 "zone_management": false, 00:08:46.967 "zone_append": false, 00:08:46.967 "compare": false, 00:08:46.967 "compare_and_write": false, 00:08:46.967 "abort": false, 00:08:46.967 "seek_hole": false, 00:08:46.967 "seek_data": false, 00:08:46.967 "copy": false, 00:08:46.967 "nvme_iov_md": false 00:08:46.967 }, 00:08:46.967 "memory_domains": [ 00:08:46.967 { 00:08:46.967 "dma_device_id": "system", 00:08:46.967 "dma_device_type": 1 00:08:46.967 }, 00:08:46.967 { 00:08:46.967 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.967 "dma_device_type": 2 00:08:46.967 }, 00:08:46.967 { 00:08:46.967 "dma_device_id": "system", 00:08:46.967 "dma_device_type": 1 00:08:46.967 }, 00:08:46.967 { 00:08:46.967 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.967 "dma_device_type": 2 00:08:46.967 }, 00:08:46.967 { 00:08:46.967 "dma_device_id": "system", 00:08:46.967 "dma_device_type": 1 00:08:46.967 }, 00:08:46.967 { 00:08:46.967 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:46.967 "dma_device_type": 2 00:08:46.967 } 00:08:46.967 ], 00:08:46.967 "driver_specific": { 00:08:46.967 "raid": { 00:08:46.967 "uuid": "b15423b5-156f-4a5c-8d6d-d43766782e25", 00:08:46.967 "strip_size_kb": 64, 00:08:46.967 "state": "online", 00:08:46.967 "raid_level": "concat", 00:08:46.967 "superblock": false, 00:08:46.967 "num_base_bdevs": 3, 00:08:46.967 "num_base_bdevs_discovered": 3, 00:08:46.967 "num_base_bdevs_operational": 3, 00:08:46.967 "base_bdevs_list": [ 00:08:46.967 { 00:08:46.967 "name": "NewBaseBdev", 00:08:46.967 "uuid": "395db821-14cf-40e8-a6f5-47ef4f3f805d", 00:08:46.967 "is_configured": true, 00:08:46.967 "data_offset": 0, 00:08:46.968 "data_size": 65536 00:08:46.968 }, 00:08:46.968 { 00:08:46.968 "name": "BaseBdev2", 00:08:46.968 "uuid": "63fbc760-dc3f-47bc-b066-8724b64a27ed", 00:08:46.968 "is_configured": true, 00:08:46.968 "data_offset": 0, 00:08:46.968 "data_size": 65536 00:08:46.968 }, 00:08:46.968 { 00:08:46.968 "name": "BaseBdev3", 00:08:46.968 "uuid": "817398c4-fd4e-43b0-bbd3-baec7642d64a", 00:08:46.968 "is_configured": true, 00:08:46.968 "data_offset": 0, 00:08:46.968 "data_size": 65536 00:08:46.968 } 00:08:46.968 ] 00:08:46.968 } 00:08:46.968 } 00:08:46.968 }' 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:46.968 BaseBdev2 00:08:46.968 BaseBdev3' 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:46.968 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.968 [2024-12-05 18:59:04.524825] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:47.228 [2024-12-05 18:59:04.524892] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:47.228 [2024-12-05 18:59:04.524961] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:47.228 [2024-12-05 18:59:04.525011] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:47.228 [2024-12-05 18:59:04.525023] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:47.228 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:47.228 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 76426 00:08:47.228 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 76426 ']' 00:08:47.228 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 76426 00:08:47.228 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:08:47.228 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:47.228 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76426 00:08:47.228 killing process with pid 76426 00:08:47.228 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:47.228 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:47.228 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76426' 00:08:47.228 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 76426 00:08:47.228 [2024-12-05 18:59:04.573121] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:47.228 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 76426 00:08:47.228 [2024-12-05 18:59:04.603440] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:47.487 18:59:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:47.487 00:08:47.487 real 0m8.507s 00:08:47.487 user 0m14.587s 00:08:47.487 sys 0m1.652s 00:08:47.487 ************************************ 00:08:47.487 END TEST raid_state_function_test 00:08:47.487 ************************************ 00:08:47.487 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.488 18:59:04 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 3 true 00:08:47.488 18:59:04 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:47.488 18:59:04 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:47.488 18:59:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:47.488 ************************************ 00:08:47.488 START TEST raid_state_function_test_sb 00:08:47.488 ************************************ 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 3 true 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=76977 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 76977' 00:08:47.488 Process raid pid: 76977 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 76977 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 76977 ']' 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:47.488 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:47.488 18:59:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:47.488 [2024-12-05 18:59:04.966946] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:08:47.488 [2024-12-05 18:59:04.967168] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:47.746 [2024-12-05 18:59:05.124306] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:47.746 [2024-12-05 18:59:05.149247] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:47.746 [2024-12-05 18:59:05.192195] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:47.746 [2024-12-05 18:59:05.192237] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:48.312 [2024-12-05 18:59:05.783292] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:48.312 [2024-12-05 18:59:05.783439] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:48.312 [2024-12-05 18:59:05.783462] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:48.312 [2024-12-05 18:59:05.783474] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:48.312 [2024-12-05 18:59:05.783480] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:48.312 [2024-12-05 18:59:05.783491] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:48.312 "name": "Existed_Raid", 00:08:48.312 "uuid": "abe78de9-5fd7-4eab-9e2b-624a88a82a60", 00:08:48.312 "strip_size_kb": 64, 00:08:48.312 "state": "configuring", 00:08:48.312 "raid_level": "concat", 00:08:48.312 "superblock": true, 00:08:48.312 "num_base_bdevs": 3, 00:08:48.312 "num_base_bdevs_discovered": 0, 00:08:48.312 "num_base_bdevs_operational": 3, 00:08:48.312 "base_bdevs_list": [ 00:08:48.312 { 00:08:48.312 "name": "BaseBdev1", 00:08:48.312 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.312 "is_configured": false, 00:08:48.312 "data_offset": 0, 00:08:48.312 "data_size": 0 00:08:48.312 }, 00:08:48.312 { 00:08:48.312 "name": "BaseBdev2", 00:08:48.312 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.312 "is_configured": false, 00:08:48.312 "data_offset": 0, 00:08:48.312 "data_size": 0 00:08:48.312 }, 00:08:48.312 { 00:08:48.312 "name": "BaseBdev3", 00:08:48.312 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.312 "is_configured": false, 00:08:48.312 "data_offset": 0, 00:08:48.312 "data_size": 0 00:08:48.312 } 00:08:48.312 ] 00:08:48.312 }' 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:48.312 18:59:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:48.879 [2024-12-05 18:59:06.234395] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:48.879 [2024-12-05 18:59:06.234488] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:48.879 [2024-12-05 18:59:06.246412] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:48.879 [2024-12-05 18:59:06.246492] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:48.879 [2024-12-05 18:59:06.246534] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:48.879 [2024-12-05 18:59:06.246556] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:48.879 [2024-12-05 18:59:06.246573] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:48.879 [2024-12-05 18:59:06.246594] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:48.879 [2024-12-05 18:59:06.267312] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:48.879 BaseBdev1 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:48.879 [ 00:08:48.879 { 00:08:48.879 "name": "BaseBdev1", 00:08:48.879 "aliases": [ 00:08:48.879 "6199dac8-aba3-4e2a-80cd-2876a29d07e4" 00:08:48.879 ], 00:08:48.879 "product_name": "Malloc disk", 00:08:48.879 "block_size": 512, 00:08:48.879 "num_blocks": 65536, 00:08:48.879 "uuid": "6199dac8-aba3-4e2a-80cd-2876a29d07e4", 00:08:48.879 "assigned_rate_limits": { 00:08:48.879 "rw_ios_per_sec": 0, 00:08:48.879 "rw_mbytes_per_sec": 0, 00:08:48.879 "r_mbytes_per_sec": 0, 00:08:48.879 "w_mbytes_per_sec": 0 00:08:48.879 }, 00:08:48.879 "claimed": true, 00:08:48.879 "claim_type": "exclusive_write", 00:08:48.879 "zoned": false, 00:08:48.879 "supported_io_types": { 00:08:48.879 "read": true, 00:08:48.879 "write": true, 00:08:48.879 "unmap": true, 00:08:48.879 "flush": true, 00:08:48.879 "reset": true, 00:08:48.879 "nvme_admin": false, 00:08:48.879 "nvme_io": false, 00:08:48.879 "nvme_io_md": false, 00:08:48.879 "write_zeroes": true, 00:08:48.879 "zcopy": true, 00:08:48.879 "get_zone_info": false, 00:08:48.879 "zone_management": false, 00:08:48.879 "zone_append": false, 00:08:48.879 "compare": false, 00:08:48.879 "compare_and_write": false, 00:08:48.879 "abort": true, 00:08:48.879 "seek_hole": false, 00:08:48.879 "seek_data": false, 00:08:48.879 "copy": true, 00:08:48.879 "nvme_iov_md": false 00:08:48.879 }, 00:08:48.879 "memory_domains": [ 00:08:48.879 { 00:08:48.879 "dma_device_id": "system", 00:08:48.879 "dma_device_type": 1 00:08:48.879 }, 00:08:48.879 { 00:08:48.879 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:48.879 "dma_device_type": 2 00:08:48.879 } 00:08:48.879 ], 00:08:48.879 "driver_specific": {} 00:08:48.879 } 00:08:48.879 ] 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:48.879 "name": "Existed_Raid", 00:08:48.879 "uuid": "3ba9bd52-d8a6-4bda-b8cb-308bcce65e1c", 00:08:48.879 "strip_size_kb": 64, 00:08:48.879 "state": "configuring", 00:08:48.879 "raid_level": "concat", 00:08:48.879 "superblock": true, 00:08:48.879 "num_base_bdevs": 3, 00:08:48.879 "num_base_bdevs_discovered": 1, 00:08:48.879 "num_base_bdevs_operational": 3, 00:08:48.879 "base_bdevs_list": [ 00:08:48.879 { 00:08:48.879 "name": "BaseBdev1", 00:08:48.879 "uuid": "6199dac8-aba3-4e2a-80cd-2876a29d07e4", 00:08:48.879 "is_configured": true, 00:08:48.879 "data_offset": 2048, 00:08:48.879 "data_size": 63488 00:08:48.879 }, 00:08:48.879 { 00:08:48.879 "name": "BaseBdev2", 00:08:48.879 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.879 "is_configured": false, 00:08:48.879 "data_offset": 0, 00:08:48.879 "data_size": 0 00:08:48.879 }, 00:08:48.879 { 00:08:48.879 "name": "BaseBdev3", 00:08:48.879 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:48.879 "is_configured": false, 00:08:48.879 "data_offset": 0, 00:08:48.879 "data_size": 0 00:08:48.879 } 00:08:48.879 ] 00:08:48.879 }' 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:48.879 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.138 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:49.138 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.138 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.138 [2024-12-05 18:59:06.694604] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:49.138 [2024-12-05 18:59:06.694698] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.398 [2024-12-05 18:59:06.706627] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:49.398 [2024-12-05 18:59:06.708526] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:49.398 [2024-12-05 18:59:06.708606] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:49.398 [2024-12-05 18:59:06.708635] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:49.398 [2024-12-05 18:59:06.708657] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:49.398 "name": "Existed_Raid", 00:08:49.398 "uuid": "0b9eee3f-558f-4f47-9a34-25e0feabf804", 00:08:49.398 "strip_size_kb": 64, 00:08:49.398 "state": "configuring", 00:08:49.398 "raid_level": "concat", 00:08:49.398 "superblock": true, 00:08:49.398 "num_base_bdevs": 3, 00:08:49.398 "num_base_bdevs_discovered": 1, 00:08:49.398 "num_base_bdevs_operational": 3, 00:08:49.398 "base_bdevs_list": [ 00:08:49.398 { 00:08:49.398 "name": "BaseBdev1", 00:08:49.398 "uuid": "6199dac8-aba3-4e2a-80cd-2876a29d07e4", 00:08:49.398 "is_configured": true, 00:08:49.398 "data_offset": 2048, 00:08:49.398 "data_size": 63488 00:08:49.398 }, 00:08:49.398 { 00:08:49.398 "name": "BaseBdev2", 00:08:49.398 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:49.398 "is_configured": false, 00:08:49.398 "data_offset": 0, 00:08:49.398 "data_size": 0 00:08:49.398 }, 00:08:49.398 { 00:08:49.398 "name": "BaseBdev3", 00:08:49.398 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:49.398 "is_configured": false, 00:08:49.398 "data_offset": 0, 00:08:49.398 "data_size": 0 00:08:49.398 } 00:08:49.398 ] 00:08:49.398 }' 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:49.398 18:59:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.658 [2024-12-05 18:59:07.160952] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:49.658 BaseBdev2 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.658 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.658 [ 00:08:49.658 { 00:08:49.658 "name": "BaseBdev2", 00:08:49.658 "aliases": [ 00:08:49.658 "e34b2878-5295-48a6-a02c-6e6ee1e514f3" 00:08:49.658 ], 00:08:49.658 "product_name": "Malloc disk", 00:08:49.658 "block_size": 512, 00:08:49.658 "num_blocks": 65536, 00:08:49.658 "uuid": "e34b2878-5295-48a6-a02c-6e6ee1e514f3", 00:08:49.658 "assigned_rate_limits": { 00:08:49.658 "rw_ios_per_sec": 0, 00:08:49.658 "rw_mbytes_per_sec": 0, 00:08:49.658 "r_mbytes_per_sec": 0, 00:08:49.658 "w_mbytes_per_sec": 0 00:08:49.658 }, 00:08:49.658 "claimed": true, 00:08:49.658 "claim_type": "exclusive_write", 00:08:49.658 "zoned": false, 00:08:49.658 "supported_io_types": { 00:08:49.658 "read": true, 00:08:49.658 "write": true, 00:08:49.658 "unmap": true, 00:08:49.658 "flush": true, 00:08:49.658 "reset": true, 00:08:49.658 "nvme_admin": false, 00:08:49.658 "nvme_io": false, 00:08:49.658 "nvme_io_md": false, 00:08:49.658 "write_zeroes": true, 00:08:49.658 "zcopy": true, 00:08:49.658 "get_zone_info": false, 00:08:49.658 "zone_management": false, 00:08:49.658 "zone_append": false, 00:08:49.658 "compare": false, 00:08:49.658 "compare_and_write": false, 00:08:49.658 "abort": true, 00:08:49.658 "seek_hole": false, 00:08:49.658 "seek_data": false, 00:08:49.658 "copy": true, 00:08:49.658 "nvme_iov_md": false 00:08:49.658 }, 00:08:49.658 "memory_domains": [ 00:08:49.658 { 00:08:49.658 "dma_device_id": "system", 00:08:49.658 "dma_device_type": 1 00:08:49.659 }, 00:08:49.659 { 00:08:49.659 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:49.659 "dma_device_type": 2 00:08:49.659 } 00:08:49.659 ], 00:08:49.659 "driver_specific": {} 00:08:49.659 } 00:08:49.659 ] 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:49.659 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:49.918 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.918 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:49.918 "name": "Existed_Raid", 00:08:49.918 "uuid": "0b9eee3f-558f-4f47-9a34-25e0feabf804", 00:08:49.918 "strip_size_kb": 64, 00:08:49.918 "state": "configuring", 00:08:49.918 "raid_level": "concat", 00:08:49.918 "superblock": true, 00:08:49.918 "num_base_bdevs": 3, 00:08:49.918 "num_base_bdevs_discovered": 2, 00:08:49.918 "num_base_bdevs_operational": 3, 00:08:49.918 "base_bdevs_list": [ 00:08:49.918 { 00:08:49.918 "name": "BaseBdev1", 00:08:49.918 "uuid": "6199dac8-aba3-4e2a-80cd-2876a29d07e4", 00:08:49.918 "is_configured": true, 00:08:49.918 "data_offset": 2048, 00:08:49.918 "data_size": 63488 00:08:49.918 }, 00:08:49.918 { 00:08:49.918 "name": "BaseBdev2", 00:08:49.918 "uuid": "e34b2878-5295-48a6-a02c-6e6ee1e514f3", 00:08:49.918 "is_configured": true, 00:08:49.918 "data_offset": 2048, 00:08:49.918 "data_size": 63488 00:08:49.918 }, 00:08:49.919 { 00:08:49.919 "name": "BaseBdev3", 00:08:49.919 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:49.919 "is_configured": false, 00:08:49.919 "data_offset": 0, 00:08:49.919 "data_size": 0 00:08:49.919 } 00:08:49.919 ] 00:08:49.919 }' 00:08:49.919 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:49.919 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.178 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:50.178 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.178 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.178 [2024-12-05 18:59:07.643714] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:50.178 [2024-12-05 18:59:07.643986] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:50.178 [2024-12-05 18:59:07.644043] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:50.178 [2024-12-05 18:59:07.644338] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:50.178 BaseBdev3 00:08:50.179 [2024-12-05 18:59:07.644569] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:50.179 [2024-12-05 18:59:07.644627] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:50.179 [2024-12-05 18:59:07.644854] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.179 [ 00:08:50.179 { 00:08:50.179 "name": "BaseBdev3", 00:08:50.179 "aliases": [ 00:08:50.179 "2544bb3b-88e2-4675-aee9-b80dc2d4dd13" 00:08:50.179 ], 00:08:50.179 "product_name": "Malloc disk", 00:08:50.179 "block_size": 512, 00:08:50.179 "num_blocks": 65536, 00:08:50.179 "uuid": "2544bb3b-88e2-4675-aee9-b80dc2d4dd13", 00:08:50.179 "assigned_rate_limits": { 00:08:50.179 "rw_ios_per_sec": 0, 00:08:50.179 "rw_mbytes_per_sec": 0, 00:08:50.179 "r_mbytes_per_sec": 0, 00:08:50.179 "w_mbytes_per_sec": 0 00:08:50.179 }, 00:08:50.179 "claimed": true, 00:08:50.179 "claim_type": "exclusive_write", 00:08:50.179 "zoned": false, 00:08:50.179 "supported_io_types": { 00:08:50.179 "read": true, 00:08:50.179 "write": true, 00:08:50.179 "unmap": true, 00:08:50.179 "flush": true, 00:08:50.179 "reset": true, 00:08:50.179 "nvme_admin": false, 00:08:50.179 "nvme_io": false, 00:08:50.179 "nvme_io_md": false, 00:08:50.179 "write_zeroes": true, 00:08:50.179 "zcopy": true, 00:08:50.179 "get_zone_info": false, 00:08:50.179 "zone_management": false, 00:08:50.179 "zone_append": false, 00:08:50.179 "compare": false, 00:08:50.179 "compare_and_write": false, 00:08:50.179 "abort": true, 00:08:50.179 "seek_hole": false, 00:08:50.179 "seek_data": false, 00:08:50.179 "copy": true, 00:08:50.179 "nvme_iov_md": false 00:08:50.179 }, 00:08:50.179 "memory_domains": [ 00:08:50.179 { 00:08:50.179 "dma_device_id": "system", 00:08:50.179 "dma_device_type": 1 00:08:50.179 }, 00:08:50.179 { 00:08:50.179 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.179 "dma_device_type": 2 00:08:50.179 } 00:08:50.179 ], 00:08:50.179 "driver_specific": {} 00:08:50.179 } 00:08:50.179 ] 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:50.179 "name": "Existed_Raid", 00:08:50.179 "uuid": "0b9eee3f-558f-4f47-9a34-25e0feabf804", 00:08:50.179 "strip_size_kb": 64, 00:08:50.179 "state": "online", 00:08:50.179 "raid_level": "concat", 00:08:50.179 "superblock": true, 00:08:50.179 "num_base_bdevs": 3, 00:08:50.179 "num_base_bdevs_discovered": 3, 00:08:50.179 "num_base_bdevs_operational": 3, 00:08:50.179 "base_bdevs_list": [ 00:08:50.179 { 00:08:50.179 "name": "BaseBdev1", 00:08:50.179 "uuid": "6199dac8-aba3-4e2a-80cd-2876a29d07e4", 00:08:50.179 "is_configured": true, 00:08:50.179 "data_offset": 2048, 00:08:50.179 "data_size": 63488 00:08:50.179 }, 00:08:50.179 { 00:08:50.179 "name": "BaseBdev2", 00:08:50.179 "uuid": "e34b2878-5295-48a6-a02c-6e6ee1e514f3", 00:08:50.179 "is_configured": true, 00:08:50.179 "data_offset": 2048, 00:08:50.179 "data_size": 63488 00:08:50.179 }, 00:08:50.179 { 00:08:50.179 "name": "BaseBdev3", 00:08:50.179 "uuid": "2544bb3b-88e2-4675-aee9-b80dc2d4dd13", 00:08:50.179 "is_configured": true, 00:08:50.179 "data_offset": 2048, 00:08:50.179 "data_size": 63488 00:08:50.179 } 00:08:50.179 ] 00:08:50.179 }' 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:50.179 18:59:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.750 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:50.750 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:50.750 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:50.750 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:50.750 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:50.750 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:50.750 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:50.750 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:50.750 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.750 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.750 [2024-12-05 18:59:08.163100] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:50.750 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:50.750 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:50.750 "name": "Existed_Raid", 00:08:50.750 "aliases": [ 00:08:50.750 "0b9eee3f-558f-4f47-9a34-25e0feabf804" 00:08:50.750 ], 00:08:50.750 "product_name": "Raid Volume", 00:08:50.750 "block_size": 512, 00:08:50.750 "num_blocks": 190464, 00:08:50.750 "uuid": "0b9eee3f-558f-4f47-9a34-25e0feabf804", 00:08:50.750 "assigned_rate_limits": { 00:08:50.750 "rw_ios_per_sec": 0, 00:08:50.750 "rw_mbytes_per_sec": 0, 00:08:50.750 "r_mbytes_per_sec": 0, 00:08:50.750 "w_mbytes_per_sec": 0 00:08:50.750 }, 00:08:50.750 "claimed": false, 00:08:50.750 "zoned": false, 00:08:50.750 "supported_io_types": { 00:08:50.750 "read": true, 00:08:50.750 "write": true, 00:08:50.750 "unmap": true, 00:08:50.750 "flush": true, 00:08:50.750 "reset": true, 00:08:50.750 "nvme_admin": false, 00:08:50.750 "nvme_io": false, 00:08:50.750 "nvme_io_md": false, 00:08:50.750 "write_zeroes": true, 00:08:50.750 "zcopy": false, 00:08:50.750 "get_zone_info": false, 00:08:50.750 "zone_management": false, 00:08:50.750 "zone_append": false, 00:08:50.750 "compare": false, 00:08:50.750 "compare_and_write": false, 00:08:50.750 "abort": false, 00:08:50.750 "seek_hole": false, 00:08:50.750 "seek_data": false, 00:08:50.750 "copy": false, 00:08:50.750 "nvme_iov_md": false 00:08:50.750 }, 00:08:50.750 "memory_domains": [ 00:08:50.750 { 00:08:50.750 "dma_device_id": "system", 00:08:50.750 "dma_device_type": 1 00:08:50.750 }, 00:08:50.750 { 00:08:50.750 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.750 "dma_device_type": 2 00:08:50.750 }, 00:08:50.750 { 00:08:50.750 "dma_device_id": "system", 00:08:50.750 "dma_device_type": 1 00:08:50.750 }, 00:08:50.750 { 00:08:50.750 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.750 "dma_device_type": 2 00:08:50.750 }, 00:08:50.750 { 00:08:50.750 "dma_device_id": "system", 00:08:50.750 "dma_device_type": 1 00:08:50.750 }, 00:08:50.750 { 00:08:50.750 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:50.750 "dma_device_type": 2 00:08:50.750 } 00:08:50.750 ], 00:08:50.750 "driver_specific": { 00:08:50.750 "raid": { 00:08:50.750 "uuid": "0b9eee3f-558f-4f47-9a34-25e0feabf804", 00:08:50.750 "strip_size_kb": 64, 00:08:50.750 "state": "online", 00:08:50.750 "raid_level": "concat", 00:08:50.750 "superblock": true, 00:08:50.750 "num_base_bdevs": 3, 00:08:50.750 "num_base_bdevs_discovered": 3, 00:08:50.750 "num_base_bdevs_operational": 3, 00:08:50.750 "base_bdevs_list": [ 00:08:50.750 { 00:08:50.750 "name": "BaseBdev1", 00:08:50.750 "uuid": "6199dac8-aba3-4e2a-80cd-2876a29d07e4", 00:08:50.750 "is_configured": true, 00:08:50.750 "data_offset": 2048, 00:08:50.750 "data_size": 63488 00:08:50.750 }, 00:08:50.750 { 00:08:50.750 "name": "BaseBdev2", 00:08:50.750 "uuid": "e34b2878-5295-48a6-a02c-6e6ee1e514f3", 00:08:50.750 "is_configured": true, 00:08:50.750 "data_offset": 2048, 00:08:50.750 "data_size": 63488 00:08:50.750 }, 00:08:50.750 { 00:08:50.750 "name": "BaseBdev3", 00:08:50.750 "uuid": "2544bb3b-88e2-4675-aee9-b80dc2d4dd13", 00:08:50.750 "is_configured": true, 00:08:50.751 "data_offset": 2048, 00:08:50.751 "data_size": 63488 00:08:50.751 } 00:08:50.751 ] 00:08:50.751 } 00:08:50.751 } 00:08:50.751 }' 00:08:50.751 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:50.751 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:50.751 BaseBdev2 00:08:50.751 BaseBdev3' 00:08:50.751 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:50.751 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:50.751 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:50.751 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:50.751 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:50.751 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:50.751 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:50.751 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.010 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:51.010 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:51.010 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:51.010 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:51.010 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:51.010 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.010 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.010 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.010 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:51.010 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:51.010 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:51.010 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.011 [2024-12-05 18:59:08.410444] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:51.011 [2024-12-05 18:59:08.410469] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:51.011 [2024-12-05 18:59:08.410528] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:51.011 "name": "Existed_Raid", 00:08:51.011 "uuid": "0b9eee3f-558f-4f47-9a34-25e0feabf804", 00:08:51.011 "strip_size_kb": 64, 00:08:51.011 "state": "offline", 00:08:51.011 "raid_level": "concat", 00:08:51.011 "superblock": true, 00:08:51.011 "num_base_bdevs": 3, 00:08:51.011 "num_base_bdevs_discovered": 2, 00:08:51.011 "num_base_bdevs_operational": 2, 00:08:51.011 "base_bdevs_list": [ 00:08:51.011 { 00:08:51.011 "name": null, 00:08:51.011 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.011 "is_configured": false, 00:08:51.011 "data_offset": 0, 00:08:51.011 "data_size": 63488 00:08:51.011 }, 00:08:51.011 { 00:08:51.011 "name": "BaseBdev2", 00:08:51.011 "uuid": "e34b2878-5295-48a6-a02c-6e6ee1e514f3", 00:08:51.011 "is_configured": true, 00:08:51.011 "data_offset": 2048, 00:08:51.011 "data_size": 63488 00:08:51.011 }, 00:08:51.011 { 00:08:51.011 "name": "BaseBdev3", 00:08:51.011 "uuid": "2544bb3b-88e2-4675-aee9-b80dc2d4dd13", 00:08:51.011 "is_configured": true, 00:08:51.011 "data_offset": 2048, 00:08:51.011 "data_size": 63488 00:08:51.011 } 00:08:51.011 ] 00:08:51.011 }' 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:51.011 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.595 [2024-12-05 18:59:08.928753] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.595 18:59:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.595 [2024-12-05 18:59:08.999751] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:51.595 [2024-12-05 18:59:08.999796] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.595 BaseBdev2 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.595 [ 00:08:51.595 { 00:08:51.595 "name": "BaseBdev2", 00:08:51.595 "aliases": [ 00:08:51.595 "b5a7afa1-f0a2-4ca6-9349-5a25e951e749" 00:08:51.595 ], 00:08:51.595 "product_name": "Malloc disk", 00:08:51.595 "block_size": 512, 00:08:51.595 "num_blocks": 65536, 00:08:51.595 "uuid": "b5a7afa1-f0a2-4ca6-9349-5a25e951e749", 00:08:51.595 "assigned_rate_limits": { 00:08:51.595 "rw_ios_per_sec": 0, 00:08:51.595 "rw_mbytes_per_sec": 0, 00:08:51.595 "r_mbytes_per_sec": 0, 00:08:51.595 "w_mbytes_per_sec": 0 00:08:51.595 }, 00:08:51.595 "claimed": false, 00:08:51.595 "zoned": false, 00:08:51.595 "supported_io_types": { 00:08:51.595 "read": true, 00:08:51.595 "write": true, 00:08:51.595 "unmap": true, 00:08:51.595 "flush": true, 00:08:51.595 "reset": true, 00:08:51.595 "nvme_admin": false, 00:08:51.595 "nvme_io": false, 00:08:51.595 "nvme_io_md": false, 00:08:51.595 "write_zeroes": true, 00:08:51.595 "zcopy": true, 00:08:51.595 "get_zone_info": false, 00:08:51.595 "zone_management": false, 00:08:51.595 "zone_append": false, 00:08:51.595 "compare": false, 00:08:51.595 "compare_and_write": false, 00:08:51.595 "abort": true, 00:08:51.595 "seek_hole": false, 00:08:51.595 "seek_data": false, 00:08:51.595 "copy": true, 00:08:51.595 "nvme_iov_md": false 00:08:51.595 }, 00:08:51.595 "memory_domains": [ 00:08:51.595 { 00:08:51.595 "dma_device_id": "system", 00:08:51.595 "dma_device_type": 1 00:08:51.595 }, 00:08:51.595 { 00:08:51.595 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:51.595 "dma_device_type": 2 00:08:51.595 } 00:08:51.595 ], 00:08:51.595 "driver_specific": {} 00:08:51.595 } 00:08:51.595 ] 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.595 BaseBdev3 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.595 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:51.596 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.596 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.885 [ 00:08:51.885 { 00:08:51.885 "name": "BaseBdev3", 00:08:51.885 "aliases": [ 00:08:51.885 "fde552f6-1462-4ebe-b031-a2d15a85c476" 00:08:51.885 ], 00:08:51.885 "product_name": "Malloc disk", 00:08:51.885 "block_size": 512, 00:08:51.885 "num_blocks": 65536, 00:08:51.885 "uuid": "fde552f6-1462-4ebe-b031-a2d15a85c476", 00:08:51.885 "assigned_rate_limits": { 00:08:51.885 "rw_ios_per_sec": 0, 00:08:51.885 "rw_mbytes_per_sec": 0, 00:08:51.885 "r_mbytes_per_sec": 0, 00:08:51.885 "w_mbytes_per_sec": 0 00:08:51.885 }, 00:08:51.885 "claimed": false, 00:08:51.885 "zoned": false, 00:08:51.885 "supported_io_types": { 00:08:51.885 "read": true, 00:08:51.885 "write": true, 00:08:51.885 "unmap": true, 00:08:51.885 "flush": true, 00:08:51.885 "reset": true, 00:08:51.885 "nvme_admin": false, 00:08:51.885 "nvme_io": false, 00:08:51.885 "nvme_io_md": false, 00:08:51.885 "write_zeroes": true, 00:08:51.885 "zcopy": true, 00:08:51.885 "get_zone_info": false, 00:08:51.885 "zone_management": false, 00:08:51.885 "zone_append": false, 00:08:51.885 "compare": false, 00:08:51.885 "compare_and_write": false, 00:08:51.885 "abort": true, 00:08:51.885 "seek_hole": false, 00:08:51.885 "seek_data": false, 00:08:51.885 "copy": true, 00:08:51.885 "nvme_iov_md": false 00:08:51.885 }, 00:08:51.885 "memory_domains": [ 00:08:51.885 { 00:08:51.885 "dma_device_id": "system", 00:08:51.885 "dma_device_type": 1 00:08:51.885 }, 00:08:51.885 { 00:08:51.885 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:51.885 "dma_device_type": 2 00:08:51.885 } 00:08:51.885 ], 00:08:51.885 "driver_specific": {} 00:08:51.885 } 00:08:51.885 ] 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.885 [2024-12-05 18:59:09.170360] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:51.885 [2024-12-05 18:59:09.170446] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:51.885 [2024-12-05 18:59:09.170503] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:51.885 [2024-12-05 18:59:09.172280] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:51.885 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:51.885 "name": "Existed_Raid", 00:08:51.885 "uuid": "f42cefb6-c524-412d-bb8b-32812c19fcf5", 00:08:51.885 "strip_size_kb": 64, 00:08:51.885 "state": "configuring", 00:08:51.886 "raid_level": "concat", 00:08:51.886 "superblock": true, 00:08:51.886 "num_base_bdevs": 3, 00:08:51.886 "num_base_bdevs_discovered": 2, 00:08:51.886 "num_base_bdevs_operational": 3, 00:08:51.886 "base_bdevs_list": [ 00:08:51.886 { 00:08:51.886 "name": "BaseBdev1", 00:08:51.886 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.886 "is_configured": false, 00:08:51.886 "data_offset": 0, 00:08:51.886 "data_size": 0 00:08:51.886 }, 00:08:51.886 { 00:08:51.886 "name": "BaseBdev2", 00:08:51.886 "uuid": "b5a7afa1-f0a2-4ca6-9349-5a25e951e749", 00:08:51.886 "is_configured": true, 00:08:51.886 "data_offset": 2048, 00:08:51.886 "data_size": 63488 00:08:51.886 }, 00:08:51.886 { 00:08:51.886 "name": "BaseBdev3", 00:08:51.886 "uuid": "fde552f6-1462-4ebe-b031-a2d15a85c476", 00:08:51.886 "is_configured": true, 00:08:51.886 "data_offset": 2048, 00:08:51.886 "data_size": 63488 00:08:51.886 } 00:08:51.886 ] 00:08:51.886 }' 00:08:51.886 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:51.886 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.146 [2024-12-05 18:59:09.637572] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:52.146 "name": "Existed_Raid", 00:08:52.146 "uuid": "f42cefb6-c524-412d-bb8b-32812c19fcf5", 00:08:52.146 "strip_size_kb": 64, 00:08:52.146 "state": "configuring", 00:08:52.146 "raid_level": "concat", 00:08:52.146 "superblock": true, 00:08:52.146 "num_base_bdevs": 3, 00:08:52.146 "num_base_bdevs_discovered": 1, 00:08:52.146 "num_base_bdevs_operational": 3, 00:08:52.146 "base_bdevs_list": [ 00:08:52.146 { 00:08:52.146 "name": "BaseBdev1", 00:08:52.146 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:52.146 "is_configured": false, 00:08:52.146 "data_offset": 0, 00:08:52.146 "data_size": 0 00:08:52.146 }, 00:08:52.146 { 00:08:52.146 "name": null, 00:08:52.146 "uuid": "b5a7afa1-f0a2-4ca6-9349-5a25e951e749", 00:08:52.146 "is_configured": false, 00:08:52.146 "data_offset": 0, 00:08:52.146 "data_size": 63488 00:08:52.146 }, 00:08:52.146 { 00:08:52.146 "name": "BaseBdev3", 00:08:52.146 "uuid": "fde552f6-1462-4ebe-b031-a2d15a85c476", 00:08:52.146 "is_configured": true, 00:08:52.146 "data_offset": 2048, 00:08:52.146 "data_size": 63488 00:08:52.146 } 00:08:52.146 ] 00:08:52.146 }' 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:52.146 18:59:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.716 [2024-12-05 18:59:10.123774] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:52.716 BaseBdev1 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.716 [ 00:08:52.716 { 00:08:52.716 "name": "BaseBdev1", 00:08:52.716 "aliases": [ 00:08:52.716 "2ab36832-9146-4552-8674-a523896727aa" 00:08:52.716 ], 00:08:52.716 "product_name": "Malloc disk", 00:08:52.716 "block_size": 512, 00:08:52.716 "num_blocks": 65536, 00:08:52.716 "uuid": "2ab36832-9146-4552-8674-a523896727aa", 00:08:52.716 "assigned_rate_limits": { 00:08:52.716 "rw_ios_per_sec": 0, 00:08:52.716 "rw_mbytes_per_sec": 0, 00:08:52.716 "r_mbytes_per_sec": 0, 00:08:52.716 "w_mbytes_per_sec": 0 00:08:52.716 }, 00:08:52.716 "claimed": true, 00:08:52.716 "claim_type": "exclusive_write", 00:08:52.716 "zoned": false, 00:08:52.716 "supported_io_types": { 00:08:52.716 "read": true, 00:08:52.716 "write": true, 00:08:52.716 "unmap": true, 00:08:52.716 "flush": true, 00:08:52.716 "reset": true, 00:08:52.716 "nvme_admin": false, 00:08:52.716 "nvme_io": false, 00:08:52.716 "nvme_io_md": false, 00:08:52.716 "write_zeroes": true, 00:08:52.716 "zcopy": true, 00:08:52.716 "get_zone_info": false, 00:08:52.716 "zone_management": false, 00:08:52.716 "zone_append": false, 00:08:52.716 "compare": false, 00:08:52.716 "compare_and_write": false, 00:08:52.716 "abort": true, 00:08:52.716 "seek_hole": false, 00:08:52.716 "seek_data": false, 00:08:52.716 "copy": true, 00:08:52.716 "nvme_iov_md": false 00:08:52.716 }, 00:08:52.716 "memory_domains": [ 00:08:52.716 { 00:08:52.716 "dma_device_id": "system", 00:08:52.716 "dma_device_type": 1 00:08:52.716 }, 00:08:52.716 { 00:08:52.716 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:52.716 "dma_device_type": 2 00:08:52.716 } 00:08:52.716 ], 00:08:52.716 "driver_specific": {} 00:08:52.716 } 00:08:52.716 ] 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:52.716 "name": "Existed_Raid", 00:08:52.716 "uuid": "f42cefb6-c524-412d-bb8b-32812c19fcf5", 00:08:52.716 "strip_size_kb": 64, 00:08:52.716 "state": "configuring", 00:08:52.716 "raid_level": "concat", 00:08:52.716 "superblock": true, 00:08:52.716 "num_base_bdevs": 3, 00:08:52.716 "num_base_bdevs_discovered": 2, 00:08:52.716 "num_base_bdevs_operational": 3, 00:08:52.716 "base_bdevs_list": [ 00:08:52.716 { 00:08:52.716 "name": "BaseBdev1", 00:08:52.716 "uuid": "2ab36832-9146-4552-8674-a523896727aa", 00:08:52.716 "is_configured": true, 00:08:52.716 "data_offset": 2048, 00:08:52.716 "data_size": 63488 00:08:52.716 }, 00:08:52.716 { 00:08:52.716 "name": null, 00:08:52.716 "uuid": "b5a7afa1-f0a2-4ca6-9349-5a25e951e749", 00:08:52.716 "is_configured": false, 00:08:52.716 "data_offset": 0, 00:08:52.716 "data_size": 63488 00:08:52.716 }, 00:08:52.716 { 00:08:52.716 "name": "BaseBdev3", 00:08:52.716 "uuid": "fde552f6-1462-4ebe-b031-a2d15a85c476", 00:08:52.716 "is_configured": true, 00:08:52.716 "data_offset": 2048, 00:08:52.716 "data_size": 63488 00:08:52.716 } 00:08:52.716 ] 00:08:52.716 }' 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:52.716 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.286 [2024-12-05 18:59:10.626958] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:53.286 "name": "Existed_Raid", 00:08:53.286 "uuid": "f42cefb6-c524-412d-bb8b-32812c19fcf5", 00:08:53.286 "strip_size_kb": 64, 00:08:53.286 "state": "configuring", 00:08:53.286 "raid_level": "concat", 00:08:53.286 "superblock": true, 00:08:53.286 "num_base_bdevs": 3, 00:08:53.286 "num_base_bdevs_discovered": 1, 00:08:53.286 "num_base_bdevs_operational": 3, 00:08:53.286 "base_bdevs_list": [ 00:08:53.286 { 00:08:53.286 "name": "BaseBdev1", 00:08:53.286 "uuid": "2ab36832-9146-4552-8674-a523896727aa", 00:08:53.286 "is_configured": true, 00:08:53.286 "data_offset": 2048, 00:08:53.286 "data_size": 63488 00:08:53.286 }, 00:08:53.286 { 00:08:53.286 "name": null, 00:08:53.286 "uuid": "b5a7afa1-f0a2-4ca6-9349-5a25e951e749", 00:08:53.286 "is_configured": false, 00:08:53.286 "data_offset": 0, 00:08:53.286 "data_size": 63488 00:08:53.286 }, 00:08:53.286 { 00:08:53.286 "name": null, 00:08:53.286 "uuid": "fde552f6-1462-4ebe-b031-a2d15a85c476", 00:08:53.286 "is_configured": false, 00:08:53.286 "data_offset": 0, 00:08:53.286 "data_size": 63488 00:08:53.286 } 00:08:53.286 ] 00:08:53.286 }' 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:53.286 18:59:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.546 [2024-12-05 18:59:11.062226] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:53.546 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:53.806 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:53.806 "name": "Existed_Raid", 00:08:53.806 "uuid": "f42cefb6-c524-412d-bb8b-32812c19fcf5", 00:08:53.806 "strip_size_kb": 64, 00:08:53.806 "state": "configuring", 00:08:53.806 "raid_level": "concat", 00:08:53.806 "superblock": true, 00:08:53.806 "num_base_bdevs": 3, 00:08:53.806 "num_base_bdevs_discovered": 2, 00:08:53.806 "num_base_bdevs_operational": 3, 00:08:53.806 "base_bdevs_list": [ 00:08:53.806 { 00:08:53.806 "name": "BaseBdev1", 00:08:53.806 "uuid": "2ab36832-9146-4552-8674-a523896727aa", 00:08:53.806 "is_configured": true, 00:08:53.806 "data_offset": 2048, 00:08:53.806 "data_size": 63488 00:08:53.806 }, 00:08:53.806 { 00:08:53.806 "name": null, 00:08:53.806 "uuid": "b5a7afa1-f0a2-4ca6-9349-5a25e951e749", 00:08:53.806 "is_configured": false, 00:08:53.806 "data_offset": 0, 00:08:53.806 "data_size": 63488 00:08:53.806 }, 00:08:53.806 { 00:08:53.806 "name": "BaseBdev3", 00:08:53.806 "uuid": "fde552f6-1462-4ebe-b031-a2d15a85c476", 00:08:53.806 "is_configured": true, 00:08:53.806 "data_offset": 2048, 00:08:53.806 "data_size": 63488 00:08:53.806 } 00:08:53.806 ] 00:08:53.806 }' 00:08:53.806 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:53.806 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.065 [2024-12-05 18:59:11.521453] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.065 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:54.065 "name": "Existed_Raid", 00:08:54.065 "uuid": "f42cefb6-c524-412d-bb8b-32812c19fcf5", 00:08:54.065 "strip_size_kb": 64, 00:08:54.065 "state": "configuring", 00:08:54.065 "raid_level": "concat", 00:08:54.065 "superblock": true, 00:08:54.065 "num_base_bdevs": 3, 00:08:54.065 "num_base_bdevs_discovered": 1, 00:08:54.065 "num_base_bdevs_operational": 3, 00:08:54.065 "base_bdevs_list": [ 00:08:54.065 { 00:08:54.065 "name": null, 00:08:54.065 "uuid": "2ab36832-9146-4552-8674-a523896727aa", 00:08:54.065 "is_configured": false, 00:08:54.065 "data_offset": 0, 00:08:54.066 "data_size": 63488 00:08:54.066 }, 00:08:54.066 { 00:08:54.066 "name": null, 00:08:54.066 "uuid": "b5a7afa1-f0a2-4ca6-9349-5a25e951e749", 00:08:54.066 "is_configured": false, 00:08:54.066 "data_offset": 0, 00:08:54.066 "data_size": 63488 00:08:54.066 }, 00:08:54.066 { 00:08:54.066 "name": "BaseBdev3", 00:08:54.066 "uuid": "fde552f6-1462-4ebe-b031-a2d15a85c476", 00:08:54.066 "is_configured": true, 00:08:54.066 "data_offset": 2048, 00:08:54.066 "data_size": 63488 00:08:54.066 } 00:08:54.066 ] 00:08:54.066 }' 00:08:54.066 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:54.066 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.634 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:54.634 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.634 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.634 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.634 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.634 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:54.634 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:54.634 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.634 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.634 [2024-12-05 18:59:11.995242] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:54.634 18:59:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.634 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:54.634 18:59:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:54.634 "name": "Existed_Raid", 00:08:54.634 "uuid": "f42cefb6-c524-412d-bb8b-32812c19fcf5", 00:08:54.634 "strip_size_kb": 64, 00:08:54.634 "state": "configuring", 00:08:54.634 "raid_level": "concat", 00:08:54.634 "superblock": true, 00:08:54.634 "num_base_bdevs": 3, 00:08:54.634 "num_base_bdevs_discovered": 2, 00:08:54.634 "num_base_bdevs_operational": 3, 00:08:54.634 "base_bdevs_list": [ 00:08:54.634 { 00:08:54.634 "name": null, 00:08:54.634 "uuid": "2ab36832-9146-4552-8674-a523896727aa", 00:08:54.634 "is_configured": false, 00:08:54.634 "data_offset": 0, 00:08:54.634 "data_size": 63488 00:08:54.634 }, 00:08:54.634 { 00:08:54.634 "name": "BaseBdev2", 00:08:54.634 "uuid": "b5a7afa1-f0a2-4ca6-9349-5a25e951e749", 00:08:54.634 "is_configured": true, 00:08:54.634 "data_offset": 2048, 00:08:54.634 "data_size": 63488 00:08:54.634 }, 00:08:54.634 { 00:08:54.634 "name": "BaseBdev3", 00:08:54.634 "uuid": "fde552f6-1462-4ebe-b031-a2d15a85c476", 00:08:54.634 "is_configured": true, 00:08:54.634 "data_offset": 2048, 00:08:54.634 "data_size": 63488 00:08:54.634 } 00:08:54.634 ] 00:08:54.634 }' 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:54.634 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.895 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.895 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:54.895 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:54.895 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:54.895 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 2ab36832-9146-4552-8674-a523896727aa 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.156 [2024-12-05 18:59:12.529228] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:55.156 [2024-12-05 18:59:12.529397] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:55.156 [2024-12-05 18:59:12.529414] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:55.156 [2024-12-05 18:59:12.529648] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:55.156 NewBaseBdev 00:08:55.156 [2024-12-05 18:59:12.529787] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:55.156 [2024-12-05 18:59:12.529797] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:55.156 [2024-12-05 18:59:12.529906] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.156 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.156 [ 00:08:55.156 { 00:08:55.156 "name": "NewBaseBdev", 00:08:55.156 "aliases": [ 00:08:55.156 "2ab36832-9146-4552-8674-a523896727aa" 00:08:55.156 ], 00:08:55.156 "product_name": "Malloc disk", 00:08:55.156 "block_size": 512, 00:08:55.156 "num_blocks": 65536, 00:08:55.156 "uuid": "2ab36832-9146-4552-8674-a523896727aa", 00:08:55.156 "assigned_rate_limits": { 00:08:55.156 "rw_ios_per_sec": 0, 00:08:55.156 "rw_mbytes_per_sec": 0, 00:08:55.156 "r_mbytes_per_sec": 0, 00:08:55.156 "w_mbytes_per_sec": 0 00:08:55.156 }, 00:08:55.156 "claimed": true, 00:08:55.156 "claim_type": "exclusive_write", 00:08:55.156 "zoned": false, 00:08:55.156 "supported_io_types": { 00:08:55.156 "read": true, 00:08:55.156 "write": true, 00:08:55.156 "unmap": true, 00:08:55.156 "flush": true, 00:08:55.156 "reset": true, 00:08:55.156 "nvme_admin": false, 00:08:55.156 "nvme_io": false, 00:08:55.156 "nvme_io_md": false, 00:08:55.156 "write_zeroes": true, 00:08:55.156 "zcopy": true, 00:08:55.156 "get_zone_info": false, 00:08:55.156 "zone_management": false, 00:08:55.157 "zone_append": false, 00:08:55.157 "compare": false, 00:08:55.157 "compare_and_write": false, 00:08:55.157 "abort": true, 00:08:55.157 "seek_hole": false, 00:08:55.157 "seek_data": false, 00:08:55.157 "copy": true, 00:08:55.157 "nvme_iov_md": false 00:08:55.157 }, 00:08:55.157 "memory_domains": [ 00:08:55.157 { 00:08:55.157 "dma_device_id": "system", 00:08:55.157 "dma_device_type": 1 00:08:55.157 }, 00:08:55.157 { 00:08:55.157 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:55.157 "dma_device_type": 2 00:08:55.157 } 00:08:55.157 ], 00:08:55.157 "driver_specific": {} 00:08:55.157 } 00:08:55.157 ] 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:55.157 "name": "Existed_Raid", 00:08:55.157 "uuid": "f42cefb6-c524-412d-bb8b-32812c19fcf5", 00:08:55.157 "strip_size_kb": 64, 00:08:55.157 "state": "online", 00:08:55.157 "raid_level": "concat", 00:08:55.157 "superblock": true, 00:08:55.157 "num_base_bdevs": 3, 00:08:55.157 "num_base_bdevs_discovered": 3, 00:08:55.157 "num_base_bdevs_operational": 3, 00:08:55.157 "base_bdevs_list": [ 00:08:55.157 { 00:08:55.157 "name": "NewBaseBdev", 00:08:55.157 "uuid": "2ab36832-9146-4552-8674-a523896727aa", 00:08:55.157 "is_configured": true, 00:08:55.157 "data_offset": 2048, 00:08:55.157 "data_size": 63488 00:08:55.157 }, 00:08:55.157 { 00:08:55.157 "name": "BaseBdev2", 00:08:55.157 "uuid": "b5a7afa1-f0a2-4ca6-9349-5a25e951e749", 00:08:55.157 "is_configured": true, 00:08:55.157 "data_offset": 2048, 00:08:55.157 "data_size": 63488 00:08:55.157 }, 00:08:55.157 { 00:08:55.157 "name": "BaseBdev3", 00:08:55.157 "uuid": "fde552f6-1462-4ebe-b031-a2d15a85c476", 00:08:55.157 "is_configured": true, 00:08:55.157 "data_offset": 2048, 00:08:55.157 "data_size": 63488 00:08:55.157 } 00:08:55.157 ] 00:08:55.157 }' 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:55.157 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.418 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:55.418 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:55.418 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:55.418 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:55.418 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:55.418 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:55.418 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:55.418 18:59:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:55.418 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.418 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.679 [2024-12-05 18:59:12.976790] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:55.679 18:59:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:55.679 "name": "Existed_Raid", 00:08:55.679 "aliases": [ 00:08:55.679 "f42cefb6-c524-412d-bb8b-32812c19fcf5" 00:08:55.679 ], 00:08:55.679 "product_name": "Raid Volume", 00:08:55.679 "block_size": 512, 00:08:55.679 "num_blocks": 190464, 00:08:55.679 "uuid": "f42cefb6-c524-412d-bb8b-32812c19fcf5", 00:08:55.679 "assigned_rate_limits": { 00:08:55.679 "rw_ios_per_sec": 0, 00:08:55.679 "rw_mbytes_per_sec": 0, 00:08:55.679 "r_mbytes_per_sec": 0, 00:08:55.679 "w_mbytes_per_sec": 0 00:08:55.679 }, 00:08:55.679 "claimed": false, 00:08:55.679 "zoned": false, 00:08:55.679 "supported_io_types": { 00:08:55.679 "read": true, 00:08:55.679 "write": true, 00:08:55.679 "unmap": true, 00:08:55.679 "flush": true, 00:08:55.679 "reset": true, 00:08:55.679 "nvme_admin": false, 00:08:55.679 "nvme_io": false, 00:08:55.679 "nvme_io_md": false, 00:08:55.679 "write_zeroes": true, 00:08:55.679 "zcopy": false, 00:08:55.679 "get_zone_info": false, 00:08:55.679 "zone_management": false, 00:08:55.679 "zone_append": false, 00:08:55.679 "compare": false, 00:08:55.679 "compare_and_write": false, 00:08:55.679 "abort": false, 00:08:55.679 "seek_hole": false, 00:08:55.679 "seek_data": false, 00:08:55.679 "copy": false, 00:08:55.679 "nvme_iov_md": false 00:08:55.679 }, 00:08:55.679 "memory_domains": [ 00:08:55.679 { 00:08:55.679 "dma_device_id": "system", 00:08:55.679 "dma_device_type": 1 00:08:55.679 }, 00:08:55.679 { 00:08:55.679 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:55.679 "dma_device_type": 2 00:08:55.679 }, 00:08:55.679 { 00:08:55.679 "dma_device_id": "system", 00:08:55.679 "dma_device_type": 1 00:08:55.679 }, 00:08:55.679 { 00:08:55.679 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:55.679 "dma_device_type": 2 00:08:55.679 }, 00:08:55.679 { 00:08:55.679 "dma_device_id": "system", 00:08:55.679 "dma_device_type": 1 00:08:55.679 }, 00:08:55.679 { 00:08:55.679 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:55.679 "dma_device_type": 2 00:08:55.679 } 00:08:55.679 ], 00:08:55.679 "driver_specific": { 00:08:55.679 "raid": { 00:08:55.679 "uuid": "f42cefb6-c524-412d-bb8b-32812c19fcf5", 00:08:55.679 "strip_size_kb": 64, 00:08:55.679 "state": "online", 00:08:55.679 "raid_level": "concat", 00:08:55.679 "superblock": true, 00:08:55.679 "num_base_bdevs": 3, 00:08:55.679 "num_base_bdevs_discovered": 3, 00:08:55.679 "num_base_bdevs_operational": 3, 00:08:55.679 "base_bdevs_list": [ 00:08:55.679 { 00:08:55.679 "name": "NewBaseBdev", 00:08:55.679 "uuid": "2ab36832-9146-4552-8674-a523896727aa", 00:08:55.679 "is_configured": true, 00:08:55.679 "data_offset": 2048, 00:08:55.679 "data_size": 63488 00:08:55.679 }, 00:08:55.679 { 00:08:55.679 "name": "BaseBdev2", 00:08:55.679 "uuid": "b5a7afa1-f0a2-4ca6-9349-5a25e951e749", 00:08:55.679 "is_configured": true, 00:08:55.679 "data_offset": 2048, 00:08:55.679 "data_size": 63488 00:08:55.679 }, 00:08:55.679 { 00:08:55.679 "name": "BaseBdev3", 00:08:55.679 "uuid": "fde552f6-1462-4ebe-b031-a2d15a85c476", 00:08:55.679 "is_configured": true, 00:08:55.679 "data_offset": 2048, 00:08:55.679 "data_size": 63488 00:08:55.679 } 00:08:55.679 ] 00:08:55.679 } 00:08:55.679 } 00:08:55.679 }' 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:55.679 BaseBdev2 00:08:55.679 BaseBdev3' 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:55.679 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:55.680 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.680 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.680 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:55.680 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:55.940 [2024-12-05 18:59:13.256018] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:55.940 [2024-12-05 18:59:13.256042] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:55.940 [2024-12-05 18:59:13.256105] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:55.940 [2024-12-05 18:59:13.256154] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:55.940 [2024-12-05 18:59:13.256166] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 76977 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 76977 ']' 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 76977 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 76977 00:08:55.940 killing process with pid 76977 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 76977' 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 76977 00:08:55.940 [2024-12-05 18:59:13.304689] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:55.940 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 76977 00:08:55.940 [2024-12-05 18:59:13.334829] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:56.200 18:59:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:56.200 00:08:56.200 real 0m8.676s 00:08:56.200 user 0m14.876s 00:08:56.200 sys 0m1.702s 00:08:56.200 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:56.200 ************************************ 00:08:56.200 END TEST raid_state_function_test_sb 00:08:56.200 ************************************ 00:08:56.200 18:59:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:56.200 18:59:13 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 3 00:08:56.200 18:59:13 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:08:56.200 18:59:13 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:56.200 18:59:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:56.200 ************************************ 00:08:56.200 START TEST raid_superblock_test 00:08:56.200 ************************************ 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 3 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=77527 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 77527 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 77527 ']' 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:56.200 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:56.200 18:59:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.200 [2024-12-05 18:59:13.704698] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:08:56.200 [2024-12-05 18:59:13.704821] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77527 ] 00:08:56.460 [2024-12-05 18:59:13.858449] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:56.460 [2024-12-05 18:59:13.883106] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:56.460 [2024-12-05 18:59:13.926297] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:56.460 [2024-12-05 18:59:13.926334] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.030 malloc1 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:57.030 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.031 [2024-12-05 18:59:14.550347] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:57.031 [2024-12-05 18:59:14.550463] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:57.031 [2024-12-05 18:59:14.550497] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:08:57.031 [2024-12-05 18:59:14.550539] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:57.031 [2024-12-05 18:59:14.552645] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:57.031 [2024-12-05 18:59:14.552738] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:57.031 pt1 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.031 malloc2 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.031 [2024-12-05 18:59:14.582962] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:57.031 [2024-12-05 18:59:14.583069] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:57.031 [2024-12-05 18:59:14.583105] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:57.031 [2024-12-05 18:59:14.583133] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:57.031 [2024-12-05 18:59:14.585201] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:57.031 [2024-12-05 18:59:14.585274] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:57.031 pt2 00:08:57.031 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.292 malloc3 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.292 [2024-12-05 18:59:14.611520] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:57.292 [2024-12-05 18:59:14.611625] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:57.292 [2024-12-05 18:59:14.611656] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:57.292 [2024-12-05 18:59:14.611720] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:57.292 [2024-12-05 18:59:14.613781] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:57.292 [2024-12-05 18:59:14.613854] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:57.292 pt3 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.292 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.292 [2024-12-05 18:59:14.623559] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:57.292 [2024-12-05 18:59:14.625460] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:57.292 [2024-12-05 18:59:14.625567] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:57.292 [2024-12-05 18:59:14.625755] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:57.292 [2024-12-05 18:59:14.625808] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:57.292 [2024-12-05 18:59:14.626088] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:57.293 [2024-12-05 18:59:14.626269] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:57.293 [2024-12-05 18:59:14.626312] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:08:57.293 [2024-12-05 18:59:14.626437] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.293 "name": "raid_bdev1", 00:08:57.293 "uuid": "9cd7691a-4839-4fc3-976d-35bb2522c47b", 00:08:57.293 "strip_size_kb": 64, 00:08:57.293 "state": "online", 00:08:57.293 "raid_level": "concat", 00:08:57.293 "superblock": true, 00:08:57.293 "num_base_bdevs": 3, 00:08:57.293 "num_base_bdevs_discovered": 3, 00:08:57.293 "num_base_bdevs_operational": 3, 00:08:57.293 "base_bdevs_list": [ 00:08:57.293 { 00:08:57.293 "name": "pt1", 00:08:57.293 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:57.293 "is_configured": true, 00:08:57.293 "data_offset": 2048, 00:08:57.293 "data_size": 63488 00:08:57.293 }, 00:08:57.293 { 00:08:57.293 "name": "pt2", 00:08:57.293 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:57.293 "is_configured": true, 00:08:57.293 "data_offset": 2048, 00:08:57.293 "data_size": 63488 00:08:57.293 }, 00:08:57.293 { 00:08:57.293 "name": "pt3", 00:08:57.293 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:57.293 "is_configured": true, 00:08:57.293 "data_offset": 2048, 00:08:57.293 "data_size": 63488 00:08:57.293 } 00:08:57.293 ] 00:08:57.293 }' 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.293 18:59:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.551 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:57.551 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:57.551 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:57.551 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:57.551 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:57.551 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:57.551 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:57.551 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:57.551 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.551 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.551 [2024-12-05 18:59:15.063080] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:57.551 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.551 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:57.551 "name": "raid_bdev1", 00:08:57.551 "aliases": [ 00:08:57.551 "9cd7691a-4839-4fc3-976d-35bb2522c47b" 00:08:57.551 ], 00:08:57.551 "product_name": "Raid Volume", 00:08:57.551 "block_size": 512, 00:08:57.551 "num_blocks": 190464, 00:08:57.551 "uuid": "9cd7691a-4839-4fc3-976d-35bb2522c47b", 00:08:57.551 "assigned_rate_limits": { 00:08:57.551 "rw_ios_per_sec": 0, 00:08:57.551 "rw_mbytes_per_sec": 0, 00:08:57.551 "r_mbytes_per_sec": 0, 00:08:57.551 "w_mbytes_per_sec": 0 00:08:57.551 }, 00:08:57.551 "claimed": false, 00:08:57.551 "zoned": false, 00:08:57.551 "supported_io_types": { 00:08:57.551 "read": true, 00:08:57.551 "write": true, 00:08:57.551 "unmap": true, 00:08:57.551 "flush": true, 00:08:57.551 "reset": true, 00:08:57.551 "nvme_admin": false, 00:08:57.551 "nvme_io": false, 00:08:57.551 "nvme_io_md": false, 00:08:57.551 "write_zeroes": true, 00:08:57.551 "zcopy": false, 00:08:57.551 "get_zone_info": false, 00:08:57.551 "zone_management": false, 00:08:57.551 "zone_append": false, 00:08:57.551 "compare": false, 00:08:57.552 "compare_and_write": false, 00:08:57.552 "abort": false, 00:08:57.552 "seek_hole": false, 00:08:57.552 "seek_data": false, 00:08:57.552 "copy": false, 00:08:57.552 "nvme_iov_md": false 00:08:57.552 }, 00:08:57.552 "memory_domains": [ 00:08:57.552 { 00:08:57.552 "dma_device_id": "system", 00:08:57.552 "dma_device_type": 1 00:08:57.552 }, 00:08:57.552 { 00:08:57.552 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:57.552 "dma_device_type": 2 00:08:57.552 }, 00:08:57.552 { 00:08:57.552 "dma_device_id": "system", 00:08:57.552 "dma_device_type": 1 00:08:57.552 }, 00:08:57.552 { 00:08:57.552 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:57.552 "dma_device_type": 2 00:08:57.552 }, 00:08:57.552 { 00:08:57.552 "dma_device_id": "system", 00:08:57.552 "dma_device_type": 1 00:08:57.552 }, 00:08:57.552 { 00:08:57.552 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:57.552 "dma_device_type": 2 00:08:57.552 } 00:08:57.552 ], 00:08:57.552 "driver_specific": { 00:08:57.552 "raid": { 00:08:57.552 "uuid": "9cd7691a-4839-4fc3-976d-35bb2522c47b", 00:08:57.552 "strip_size_kb": 64, 00:08:57.552 "state": "online", 00:08:57.552 "raid_level": "concat", 00:08:57.552 "superblock": true, 00:08:57.552 "num_base_bdevs": 3, 00:08:57.552 "num_base_bdevs_discovered": 3, 00:08:57.552 "num_base_bdevs_operational": 3, 00:08:57.552 "base_bdevs_list": [ 00:08:57.552 { 00:08:57.552 "name": "pt1", 00:08:57.552 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:57.552 "is_configured": true, 00:08:57.552 "data_offset": 2048, 00:08:57.552 "data_size": 63488 00:08:57.552 }, 00:08:57.552 { 00:08:57.552 "name": "pt2", 00:08:57.552 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:57.552 "is_configured": true, 00:08:57.552 "data_offset": 2048, 00:08:57.552 "data_size": 63488 00:08:57.552 }, 00:08:57.552 { 00:08:57.552 "name": "pt3", 00:08:57.552 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:57.552 "is_configured": true, 00:08:57.552 "data_offset": 2048, 00:08:57.552 "data_size": 63488 00:08:57.552 } 00:08:57.552 ] 00:08:57.552 } 00:08:57.552 } 00:08:57.552 }' 00:08:57.552 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:57.552 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:57.552 pt2 00:08:57.552 pt3' 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.810 [2024-12-05 18:59:15.318607] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=9cd7691a-4839-4fc3-976d-35bb2522c47b 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 9cd7691a-4839-4fc3-976d-35bb2522c47b ']' 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:57.810 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.810 [2024-12-05 18:59:15.366274] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:57.810 [2024-12-05 18:59:15.366338] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:57.810 [2024-12-05 18:59:15.366420] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:57.810 [2024-12-05 18:59:15.366500] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:57.810 [2024-12-05 18:59:15.366562] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.070 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.070 [2024-12-05 18:59:15.514068] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:58.070 [2024-12-05 18:59:15.516027] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:58.070 [2024-12-05 18:59:15.516070] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:08:58.070 [2024-12-05 18:59:15.516116] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:58.070 [2024-12-05 18:59:15.516170] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:58.070 [2024-12-05 18:59:15.516201] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:08:58.070 [2024-12-05 18:59:15.516213] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:58.070 [2024-12-05 18:59:15.516232] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:08:58.070 request: 00:08:58.070 { 00:08:58.070 "name": "raid_bdev1", 00:08:58.070 "raid_level": "concat", 00:08:58.070 "base_bdevs": [ 00:08:58.070 "malloc1", 00:08:58.070 "malloc2", 00:08:58.070 "malloc3" 00:08:58.070 ], 00:08:58.070 "strip_size_kb": 64, 00:08:58.070 "superblock": false, 00:08:58.070 "method": "bdev_raid_create", 00:08:58.070 "req_id": 1 00:08:58.070 } 00:08:58.070 Got JSON-RPC error response 00:08:58.071 response: 00:08:58.071 { 00:08:58.071 "code": -17, 00:08:58.071 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:58.071 } 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.071 [2024-12-05 18:59:15.577907] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:58.071 [2024-12-05 18:59:15.577957] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:58.071 [2024-12-05 18:59:15.577973] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:58.071 [2024-12-05 18:59:15.577983] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:58.071 [2024-12-05 18:59:15.580119] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:58.071 [2024-12-05 18:59:15.580160] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:58.071 [2024-12-05 18:59:15.580221] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:58.071 [2024-12-05 18:59:15.580268] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:58.071 pt1 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.071 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.330 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:58.330 "name": "raid_bdev1", 00:08:58.330 "uuid": "9cd7691a-4839-4fc3-976d-35bb2522c47b", 00:08:58.330 "strip_size_kb": 64, 00:08:58.330 "state": "configuring", 00:08:58.330 "raid_level": "concat", 00:08:58.330 "superblock": true, 00:08:58.330 "num_base_bdevs": 3, 00:08:58.330 "num_base_bdevs_discovered": 1, 00:08:58.330 "num_base_bdevs_operational": 3, 00:08:58.330 "base_bdevs_list": [ 00:08:58.330 { 00:08:58.330 "name": "pt1", 00:08:58.330 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:58.330 "is_configured": true, 00:08:58.330 "data_offset": 2048, 00:08:58.330 "data_size": 63488 00:08:58.330 }, 00:08:58.330 { 00:08:58.330 "name": null, 00:08:58.330 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:58.330 "is_configured": false, 00:08:58.330 "data_offset": 2048, 00:08:58.330 "data_size": 63488 00:08:58.330 }, 00:08:58.330 { 00:08:58.330 "name": null, 00:08:58.330 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:58.330 "is_configured": false, 00:08:58.330 "data_offset": 2048, 00:08:58.330 "data_size": 63488 00:08:58.330 } 00:08:58.330 ] 00:08:58.330 }' 00:08:58.330 18:59:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:58.330 18:59:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.589 [2024-12-05 18:59:16.017148] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:58.589 [2024-12-05 18:59:16.017249] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:58.589 [2024-12-05 18:59:16.017286] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:08:58.589 [2024-12-05 18:59:16.017317] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:58.589 [2024-12-05 18:59:16.017697] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:58.589 [2024-12-05 18:59:16.017755] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:58.589 [2024-12-05 18:59:16.017846] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:58.589 [2024-12-05 18:59:16.017895] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:58.589 pt2 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.589 [2024-12-05 18:59:16.029140] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:58.589 "name": "raid_bdev1", 00:08:58.589 "uuid": "9cd7691a-4839-4fc3-976d-35bb2522c47b", 00:08:58.589 "strip_size_kb": 64, 00:08:58.589 "state": "configuring", 00:08:58.589 "raid_level": "concat", 00:08:58.589 "superblock": true, 00:08:58.589 "num_base_bdevs": 3, 00:08:58.589 "num_base_bdevs_discovered": 1, 00:08:58.589 "num_base_bdevs_operational": 3, 00:08:58.589 "base_bdevs_list": [ 00:08:58.589 { 00:08:58.589 "name": "pt1", 00:08:58.589 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:58.589 "is_configured": true, 00:08:58.589 "data_offset": 2048, 00:08:58.589 "data_size": 63488 00:08:58.589 }, 00:08:58.589 { 00:08:58.589 "name": null, 00:08:58.589 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:58.589 "is_configured": false, 00:08:58.589 "data_offset": 0, 00:08:58.589 "data_size": 63488 00:08:58.589 }, 00:08:58.589 { 00:08:58.589 "name": null, 00:08:58.589 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:58.589 "is_configured": false, 00:08:58.589 "data_offset": 2048, 00:08:58.589 "data_size": 63488 00:08:58.589 } 00:08:58.589 ] 00:08:58.589 }' 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:58.589 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.156 [2024-12-05 18:59:16.452418] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:59.156 [2024-12-05 18:59:16.452470] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:59.156 [2024-12-05 18:59:16.452486] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:59.156 [2024-12-05 18:59:16.452494] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:59.156 [2024-12-05 18:59:16.452850] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:59.156 [2024-12-05 18:59:16.452867] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:59.156 [2024-12-05 18:59:16.452926] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:59.156 [2024-12-05 18:59:16.452944] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:59.156 pt2 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.156 [2024-12-05 18:59:16.460412] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:59.156 [2024-12-05 18:59:16.460457] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:59.156 [2024-12-05 18:59:16.460473] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:08:59.156 [2024-12-05 18:59:16.460481] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:59.156 [2024-12-05 18:59:16.460799] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:59.156 [2024-12-05 18:59:16.460815] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:59.156 [2024-12-05 18:59:16.460865] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:59.156 [2024-12-05 18:59:16.460881] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:59.156 [2024-12-05 18:59:16.460968] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:59.156 [2024-12-05 18:59:16.460976] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:59.156 [2024-12-05 18:59:16.461187] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:59.156 [2024-12-05 18:59:16.461294] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:59.156 [2024-12-05 18:59:16.461306] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:59.156 [2024-12-05 18:59:16.461396] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:59.156 pt3 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:59.156 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.157 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:59.157 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.157 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.157 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.157 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:59.157 "name": "raid_bdev1", 00:08:59.157 "uuid": "9cd7691a-4839-4fc3-976d-35bb2522c47b", 00:08:59.157 "strip_size_kb": 64, 00:08:59.157 "state": "online", 00:08:59.157 "raid_level": "concat", 00:08:59.157 "superblock": true, 00:08:59.157 "num_base_bdevs": 3, 00:08:59.157 "num_base_bdevs_discovered": 3, 00:08:59.157 "num_base_bdevs_operational": 3, 00:08:59.157 "base_bdevs_list": [ 00:08:59.157 { 00:08:59.157 "name": "pt1", 00:08:59.157 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:59.157 "is_configured": true, 00:08:59.157 "data_offset": 2048, 00:08:59.157 "data_size": 63488 00:08:59.157 }, 00:08:59.157 { 00:08:59.157 "name": "pt2", 00:08:59.157 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:59.157 "is_configured": true, 00:08:59.157 "data_offset": 2048, 00:08:59.157 "data_size": 63488 00:08:59.157 }, 00:08:59.157 { 00:08:59.157 "name": "pt3", 00:08:59.157 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:59.157 "is_configured": true, 00:08:59.157 "data_offset": 2048, 00:08:59.157 "data_size": 63488 00:08:59.157 } 00:08:59.157 ] 00:08:59.157 }' 00:08:59.157 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:59.157 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.416 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:59.416 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:59.416 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:59.416 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:59.416 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:59.416 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:59.416 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:59.416 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:59.416 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.416 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.416 [2024-12-05 18:59:16.876072] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:59.416 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.416 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:59.416 "name": "raid_bdev1", 00:08:59.416 "aliases": [ 00:08:59.416 "9cd7691a-4839-4fc3-976d-35bb2522c47b" 00:08:59.416 ], 00:08:59.416 "product_name": "Raid Volume", 00:08:59.416 "block_size": 512, 00:08:59.416 "num_blocks": 190464, 00:08:59.416 "uuid": "9cd7691a-4839-4fc3-976d-35bb2522c47b", 00:08:59.416 "assigned_rate_limits": { 00:08:59.416 "rw_ios_per_sec": 0, 00:08:59.416 "rw_mbytes_per_sec": 0, 00:08:59.416 "r_mbytes_per_sec": 0, 00:08:59.416 "w_mbytes_per_sec": 0 00:08:59.416 }, 00:08:59.416 "claimed": false, 00:08:59.416 "zoned": false, 00:08:59.416 "supported_io_types": { 00:08:59.416 "read": true, 00:08:59.416 "write": true, 00:08:59.416 "unmap": true, 00:08:59.416 "flush": true, 00:08:59.416 "reset": true, 00:08:59.416 "nvme_admin": false, 00:08:59.416 "nvme_io": false, 00:08:59.416 "nvme_io_md": false, 00:08:59.416 "write_zeroes": true, 00:08:59.416 "zcopy": false, 00:08:59.416 "get_zone_info": false, 00:08:59.416 "zone_management": false, 00:08:59.416 "zone_append": false, 00:08:59.416 "compare": false, 00:08:59.416 "compare_and_write": false, 00:08:59.416 "abort": false, 00:08:59.416 "seek_hole": false, 00:08:59.416 "seek_data": false, 00:08:59.416 "copy": false, 00:08:59.416 "nvme_iov_md": false 00:08:59.416 }, 00:08:59.416 "memory_domains": [ 00:08:59.416 { 00:08:59.416 "dma_device_id": "system", 00:08:59.416 "dma_device_type": 1 00:08:59.416 }, 00:08:59.416 { 00:08:59.416 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:59.416 "dma_device_type": 2 00:08:59.416 }, 00:08:59.416 { 00:08:59.416 "dma_device_id": "system", 00:08:59.416 "dma_device_type": 1 00:08:59.416 }, 00:08:59.416 { 00:08:59.416 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:59.416 "dma_device_type": 2 00:08:59.416 }, 00:08:59.416 { 00:08:59.416 "dma_device_id": "system", 00:08:59.416 "dma_device_type": 1 00:08:59.416 }, 00:08:59.416 { 00:08:59.416 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:59.416 "dma_device_type": 2 00:08:59.416 } 00:08:59.416 ], 00:08:59.416 "driver_specific": { 00:08:59.416 "raid": { 00:08:59.416 "uuid": "9cd7691a-4839-4fc3-976d-35bb2522c47b", 00:08:59.416 "strip_size_kb": 64, 00:08:59.416 "state": "online", 00:08:59.416 "raid_level": "concat", 00:08:59.416 "superblock": true, 00:08:59.416 "num_base_bdevs": 3, 00:08:59.416 "num_base_bdevs_discovered": 3, 00:08:59.416 "num_base_bdevs_operational": 3, 00:08:59.416 "base_bdevs_list": [ 00:08:59.416 { 00:08:59.416 "name": "pt1", 00:08:59.416 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:59.416 "is_configured": true, 00:08:59.416 "data_offset": 2048, 00:08:59.416 "data_size": 63488 00:08:59.416 }, 00:08:59.416 { 00:08:59.416 "name": "pt2", 00:08:59.416 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:59.416 "is_configured": true, 00:08:59.416 "data_offset": 2048, 00:08:59.416 "data_size": 63488 00:08:59.416 }, 00:08:59.416 { 00:08:59.416 "name": "pt3", 00:08:59.416 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:59.416 "is_configured": true, 00:08:59.416 "data_offset": 2048, 00:08:59.416 "data_size": 63488 00:08:59.416 } 00:08:59.416 ] 00:08:59.416 } 00:08:59.416 } 00:08:59.416 }' 00:08:59.416 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:59.416 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:59.416 pt2 00:08:59.416 pt3' 00:08:59.416 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:59.675 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:59.675 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:59.675 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:59.675 18:59:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:59.675 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.676 18:59:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.676 [2024-12-05 18:59:17.119600] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 9cd7691a-4839-4fc3-976d-35bb2522c47b '!=' 9cd7691a-4839-4fc3-976d-35bb2522c47b ']' 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 77527 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 77527 ']' 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 77527 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77527 00:08:59.676 killing process with pid 77527 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77527' 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 77527 00:08:59.676 [2024-12-05 18:59:17.198836] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:59.676 [2024-12-05 18:59:17.198906] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:59.676 [2024-12-05 18:59:17.198969] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:59.676 [2024-12-05 18:59:17.198978] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:59.676 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 77527 00:08:59.676 [2024-12-05 18:59:17.231436] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:59.935 18:59:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:59.935 00:08:59.935 real 0m3.823s 00:08:59.935 user 0m6.062s 00:08:59.935 sys 0m0.803s 00:08:59.935 ************************************ 00:08:59.935 END TEST raid_superblock_test 00:08:59.935 ************************************ 00:08:59.935 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:59.935 18:59:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.193 18:59:17 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 3 read 00:09:00.193 18:59:17 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:00.193 18:59:17 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:00.193 18:59:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:00.193 ************************************ 00:09:00.193 START TEST raid_read_error_test 00:09:00.193 ************************************ 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 3 read 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.WbdfOek2X2 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=77740 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 77740 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 77740 ']' 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:00.193 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:00.193 18:59:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.193 [2024-12-05 18:59:17.617303] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:09:00.193 [2024-12-05 18:59:17.617520] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77740 ] 00:09:00.451 [2024-12-05 18:59:17.771460] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:00.451 [2024-12-05 18:59:17.796284] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:00.451 [2024-12-05 18:59:17.839256] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:00.451 [2024-12-05 18:59:17.839292] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.019 BaseBdev1_malloc 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.019 true 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.019 [2024-12-05 18:59:18.467179] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:01.019 [2024-12-05 18:59:18.467237] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:01.019 [2024-12-05 18:59:18.467273] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:01.019 [2024-12-05 18:59:18.467281] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:01.019 [2024-12-05 18:59:18.469438] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:01.019 [2024-12-05 18:59:18.469474] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:01.019 BaseBdev1 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.019 BaseBdev2_malloc 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.019 true 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.019 [2024-12-05 18:59:18.507830] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:01.019 [2024-12-05 18:59:18.507878] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:01.019 [2024-12-05 18:59:18.507896] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:01.019 [2024-12-05 18:59:18.507911] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:01.019 [2024-12-05 18:59:18.509972] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:01.019 [2024-12-05 18:59:18.510050] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:01.019 BaseBdev2 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.019 BaseBdev3_malloc 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.019 true 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.019 [2024-12-05 18:59:18.548459] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:01.019 [2024-12-05 18:59:18.548505] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:01.019 [2024-12-05 18:59:18.548523] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:01.019 [2024-12-05 18:59:18.548531] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:01.019 [2024-12-05 18:59:18.550664] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:01.019 [2024-12-05 18:59:18.550711] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:01.019 BaseBdev3 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.019 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.019 [2024-12-05 18:59:18.560487] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:01.019 [2024-12-05 18:59:18.562339] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:01.019 [2024-12-05 18:59:18.562410] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:01.019 [2024-12-05 18:59:18.562589] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:01.020 [2024-12-05 18:59:18.562604] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:01.020 [2024-12-05 18:59:18.562902] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:09:01.020 [2024-12-05 18:59:18.563064] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:01.020 [2024-12-05 18:59:18.563095] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:01.020 [2024-12-05 18:59:18.563222] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:01.020 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.020 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:01.020 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:01.020 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:01.020 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:01.020 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:01.020 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:01.020 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.020 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.020 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.020 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.020 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.020 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:01.020 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:01.020 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.277 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:01.277 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:01.277 "name": "raid_bdev1", 00:09:01.277 "uuid": "b4143543-4320-404b-a43b-2b461cfd8421", 00:09:01.277 "strip_size_kb": 64, 00:09:01.277 "state": "online", 00:09:01.277 "raid_level": "concat", 00:09:01.277 "superblock": true, 00:09:01.277 "num_base_bdevs": 3, 00:09:01.277 "num_base_bdevs_discovered": 3, 00:09:01.277 "num_base_bdevs_operational": 3, 00:09:01.277 "base_bdevs_list": [ 00:09:01.277 { 00:09:01.277 "name": "BaseBdev1", 00:09:01.277 "uuid": "5c8482dc-46ab-58c1-8336-acfa78472818", 00:09:01.277 "is_configured": true, 00:09:01.277 "data_offset": 2048, 00:09:01.277 "data_size": 63488 00:09:01.277 }, 00:09:01.277 { 00:09:01.277 "name": "BaseBdev2", 00:09:01.277 "uuid": "157288b2-be0d-5813-a647-3bbc778fb264", 00:09:01.277 "is_configured": true, 00:09:01.277 "data_offset": 2048, 00:09:01.277 "data_size": 63488 00:09:01.277 }, 00:09:01.277 { 00:09:01.277 "name": "BaseBdev3", 00:09:01.277 "uuid": "e8df4aa2-85d7-5fb9-871a-f408a6fe71e8", 00:09:01.277 "is_configured": true, 00:09:01.277 "data_offset": 2048, 00:09:01.277 "data_size": 63488 00:09:01.277 } 00:09:01.277 ] 00:09:01.277 }' 00:09:01.277 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:01.277 18:59:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.536 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:01.536 18:59:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:01.536 [2024-12-05 18:59:19.088058] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.471 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.730 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.730 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.730 "name": "raid_bdev1", 00:09:02.730 "uuid": "b4143543-4320-404b-a43b-2b461cfd8421", 00:09:02.730 "strip_size_kb": 64, 00:09:02.730 "state": "online", 00:09:02.730 "raid_level": "concat", 00:09:02.730 "superblock": true, 00:09:02.730 "num_base_bdevs": 3, 00:09:02.730 "num_base_bdevs_discovered": 3, 00:09:02.730 "num_base_bdevs_operational": 3, 00:09:02.730 "base_bdevs_list": [ 00:09:02.730 { 00:09:02.730 "name": "BaseBdev1", 00:09:02.730 "uuid": "5c8482dc-46ab-58c1-8336-acfa78472818", 00:09:02.730 "is_configured": true, 00:09:02.730 "data_offset": 2048, 00:09:02.730 "data_size": 63488 00:09:02.730 }, 00:09:02.730 { 00:09:02.730 "name": "BaseBdev2", 00:09:02.730 "uuid": "157288b2-be0d-5813-a647-3bbc778fb264", 00:09:02.730 "is_configured": true, 00:09:02.730 "data_offset": 2048, 00:09:02.730 "data_size": 63488 00:09:02.730 }, 00:09:02.730 { 00:09:02.730 "name": "BaseBdev3", 00:09:02.730 "uuid": "e8df4aa2-85d7-5fb9-871a-f408a6fe71e8", 00:09:02.730 "is_configured": true, 00:09:02.730 "data_offset": 2048, 00:09:02.730 "data_size": 63488 00:09:02.730 } 00:09:02.730 ] 00:09:02.730 }' 00:09:02.730 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.730 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.990 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:02.990 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:02.990 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.990 [2024-12-05 18:59:20.447865] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:02.990 [2024-12-05 18:59:20.447902] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:02.990 [2024-12-05 18:59:20.450423] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:02.990 [2024-12-05 18:59:20.450478] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:02.990 [2024-12-05 18:59:20.450514] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:02.990 [2024-12-05 18:59:20.450531] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:02.990 { 00:09:02.990 "results": [ 00:09:02.990 { 00:09:02.990 "job": "raid_bdev1", 00:09:02.990 "core_mask": "0x1", 00:09:02.990 "workload": "randrw", 00:09:02.990 "percentage": 50, 00:09:02.990 "status": "finished", 00:09:02.990 "queue_depth": 1, 00:09:02.990 "io_size": 131072, 00:09:02.990 "runtime": 1.360701, 00:09:02.990 "iops": 17153.65829818601, 00:09:02.990 "mibps": 2144.207287273251, 00:09:02.990 "io_failed": 1, 00:09:02.990 "io_timeout": 0, 00:09:02.990 "avg_latency_us": 80.47009843006533, 00:09:02.990 "min_latency_us": 24.929257641921396, 00:09:02.990 "max_latency_us": 1359.3711790393013 00:09:02.990 } 00:09:02.990 ], 00:09:02.990 "core_count": 1 00:09:02.990 } 00:09:02.990 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:02.990 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 77740 00:09:02.990 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 77740 ']' 00:09:02.990 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 77740 00:09:02.990 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:09:02.990 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:02.990 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77740 00:09:02.990 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:02.990 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:02.990 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77740' 00:09:02.990 killing process with pid 77740 00:09:02.990 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 77740 00:09:02.990 [2024-12-05 18:59:20.498334] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:02.990 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 77740 00:09:02.990 [2024-12-05 18:59:20.523310] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:03.257 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.WbdfOek2X2 00:09:03.257 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:03.257 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:03.257 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:09:03.257 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:09:03.257 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:03.257 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:03.257 18:59:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:09:03.257 ************************************ 00:09:03.257 END TEST raid_read_error_test 00:09:03.257 ************************************ 00:09:03.257 00:09:03.257 real 0m3.223s 00:09:03.257 user 0m4.091s 00:09:03.257 sys 0m0.491s 00:09:03.257 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:03.257 18:59:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.257 18:59:20 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 3 write 00:09:03.257 18:59:20 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:03.257 18:59:20 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:03.257 18:59:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:03.527 ************************************ 00:09:03.527 START TEST raid_write_error_test 00:09:03.527 ************************************ 00:09:03.527 18:59:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 3 write 00:09:03.527 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:09:03.527 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:03.527 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:03.527 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:03.527 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:03.527 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:03.527 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.uWT6XDO46F 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=77851 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 77851 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 77851 ']' 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:03.528 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:03.528 18:59:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.528 [2024-12-05 18:59:20.913751] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:09:03.528 [2024-12-05 18:59:20.913857] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77851 ] 00:09:03.528 [2024-12-05 18:59:21.067429] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:03.787 [2024-12-05 18:59:21.092248] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:03.787 [2024-12-05 18:59:21.135311] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:03.787 [2024-12-05 18:59:21.135436] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.383 BaseBdev1_malloc 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.383 true 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.383 [2024-12-05 18:59:21.763502] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:04.383 [2024-12-05 18:59:21.763557] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:04.383 [2024-12-05 18:59:21.763592] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:04.383 [2024-12-05 18:59:21.763601] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:04.383 [2024-12-05 18:59:21.765768] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:04.383 [2024-12-05 18:59:21.765802] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:04.383 BaseBdev1 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.383 BaseBdev2_malloc 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.383 true 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.383 [2024-12-05 18:59:21.804136] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:04.383 [2024-12-05 18:59:21.804186] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:04.383 [2024-12-05 18:59:21.804219] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:04.383 [2024-12-05 18:59:21.804235] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:04.383 [2024-12-05 18:59:21.806274] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:04.383 [2024-12-05 18:59:21.806313] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:04.383 BaseBdev2 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.383 BaseBdev3_malloc 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.383 true 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.383 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.383 [2024-12-05 18:59:21.844710] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:04.383 [2024-12-05 18:59:21.844756] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:04.383 [2024-12-05 18:59:21.844775] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:04.383 [2024-12-05 18:59:21.844783] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:04.383 [2024-12-05 18:59:21.846801] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:04.383 [2024-12-05 18:59:21.846896] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:04.384 BaseBdev3 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.384 [2024-12-05 18:59:21.856745] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:04.384 [2024-12-05 18:59:21.858552] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:04.384 [2024-12-05 18:59:21.858620] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:04.384 [2024-12-05 18:59:21.858822] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:04.384 [2024-12-05 18:59:21.858837] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:04.384 [2024-12-05 18:59:21.859085] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:09:04.384 [2024-12-05 18:59:21.859245] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:04.384 [2024-12-05 18:59:21.859256] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:04.384 [2024-12-05 18:59:21.859377] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:04.384 "name": "raid_bdev1", 00:09:04.384 "uuid": "7dbe1b59-cb77-45fd-b7b4-f90fd9eb2aa7", 00:09:04.384 "strip_size_kb": 64, 00:09:04.384 "state": "online", 00:09:04.384 "raid_level": "concat", 00:09:04.384 "superblock": true, 00:09:04.384 "num_base_bdevs": 3, 00:09:04.384 "num_base_bdevs_discovered": 3, 00:09:04.384 "num_base_bdevs_operational": 3, 00:09:04.384 "base_bdevs_list": [ 00:09:04.384 { 00:09:04.384 "name": "BaseBdev1", 00:09:04.384 "uuid": "467c9785-076c-5f7c-b110-2a180119313f", 00:09:04.384 "is_configured": true, 00:09:04.384 "data_offset": 2048, 00:09:04.384 "data_size": 63488 00:09:04.384 }, 00:09:04.384 { 00:09:04.384 "name": "BaseBdev2", 00:09:04.384 "uuid": "bbcc992c-5cf1-54cf-950c-b9ea45384b99", 00:09:04.384 "is_configured": true, 00:09:04.384 "data_offset": 2048, 00:09:04.384 "data_size": 63488 00:09:04.384 }, 00:09:04.384 { 00:09:04.384 "name": "BaseBdev3", 00:09:04.384 "uuid": "fbc331d6-2ea5-54fa-93c6-43641befc036", 00:09:04.384 "is_configured": true, 00:09:04.384 "data_offset": 2048, 00:09:04.384 "data_size": 63488 00:09:04.384 } 00:09:04.384 ] 00:09:04.384 }' 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:04.384 18:59:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.955 18:59:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:04.955 18:59:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:04.955 [2024-12-05 18:59:22.372230] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.896 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.897 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:05.897 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:05.897 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:05.897 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:05.897 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.897 "name": "raid_bdev1", 00:09:05.897 "uuid": "7dbe1b59-cb77-45fd-b7b4-f90fd9eb2aa7", 00:09:05.897 "strip_size_kb": 64, 00:09:05.897 "state": "online", 00:09:05.897 "raid_level": "concat", 00:09:05.897 "superblock": true, 00:09:05.897 "num_base_bdevs": 3, 00:09:05.897 "num_base_bdevs_discovered": 3, 00:09:05.897 "num_base_bdevs_operational": 3, 00:09:05.897 "base_bdevs_list": [ 00:09:05.897 { 00:09:05.897 "name": "BaseBdev1", 00:09:05.897 "uuid": "467c9785-076c-5f7c-b110-2a180119313f", 00:09:05.897 "is_configured": true, 00:09:05.897 "data_offset": 2048, 00:09:05.897 "data_size": 63488 00:09:05.897 }, 00:09:05.897 { 00:09:05.897 "name": "BaseBdev2", 00:09:05.897 "uuid": "bbcc992c-5cf1-54cf-950c-b9ea45384b99", 00:09:05.897 "is_configured": true, 00:09:05.897 "data_offset": 2048, 00:09:05.897 "data_size": 63488 00:09:05.897 }, 00:09:05.897 { 00:09:05.897 "name": "BaseBdev3", 00:09:05.897 "uuid": "fbc331d6-2ea5-54fa-93c6-43641befc036", 00:09:05.897 "is_configured": true, 00:09:05.897 "data_offset": 2048, 00:09:05.897 "data_size": 63488 00:09:05.897 } 00:09:05.897 ] 00:09:05.897 }' 00:09:05.897 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.897 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.467 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:06.467 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:06.467 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.467 [2024-12-05 18:59:23.796081] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:06.467 [2024-12-05 18:59:23.796194] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:06.467 [2024-12-05 18:59:23.798704] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:06.467 [2024-12-05 18:59:23.798797] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:06.467 [2024-12-05 18:59:23.798850] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:06.467 [2024-12-05 18:59:23.798906] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:06.467 { 00:09:06.467 "results": [ 00:09:06.467 { 00:09:06.467 "job": "raid_bdev1", 00:09:06.467 "core_mask": "0x1", 00:09:06.467 "workload": "randrw", 00:09:06.467 "percentage": 50, 00:09:06.467 "status": "finished", 00:09:06.467 "queue_depth": 1, 00:09:06.467 "io_size": 131072, 00:09:06.467 "runtime": 1.424953, 00:09:06.467 "iops": 17103.72201749812, 00:09:06.467 "mibps": 2137.965252187265, 00:09:06.467 "io_failed": 1, 00:09:06.467 "io_timeout": 0, 00:09:06.467 "avg_latency_us": 80.76721047719603, 00:09:06.467 "min_latency_us": 24.929257641921396, 00:09:06.467 "max_latency_us": 1345.0620087336245 00:09:06.467 } 00:09:06.467 ], 00:09:06.467 "core_count": 1 00:09:06.467 } 00:09:06.467 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:06.467 18:59:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 77851 00:09:06.467 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 77851 ']' 00:09:06.467 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 77851 00:09:06.467 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:09:06.467 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:06.467 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77851 00:09:06.467 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:06.467 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:06.467 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77851' 00:09:06.467 killing process with pid 77851 00:09:06.467 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 77851 00:09:06.467 [2024-12-05 18:59:23.841656] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:06.467 18:59:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 77851 00:09:06.467 [2024-12-05 18:59:23.866181] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:06.728 18:59:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.uWT6XDO46F 00:09:06.728 18:59:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:06.728 18:59:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:06.728 18:59:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.70 00:09:06.728 18:59:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:09:06.728 18:59:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:06.728 18:59:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:06.728 18:59:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.70 != \0\.\0\0 ]] 00:09:06.728 00:09:06.728 real 0m3.265s 00:09:06.728 user 0m4.168s 00:09:06.728 sys 0m0.519s 00:09:06.728 18:59:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:06.728 ************************************ 00:09:06.728 END TEST raid_write_error_test 00:09:06.728 ************************************ 00:09:06.728 18:59:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.728 18:59:24 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:06.728 18:59:24 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 3 false 00:09:06.728 18:59:24 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:06.728 18:59:24 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:06.728 18:59:24 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:06.728 ************************************ 00:09:06.728 START TEST raid_state_function_test 00:09:06.728 ************************************ 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 3 false 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=77965 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 77965' 00:09:06.728 Process raid pid: 77965 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 77965 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 77965 ']' 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:06.728 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:06.728 18:59:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:06.728 [2024-12-05 18:59:24.238328] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:09:06.728 [2024-12-05 18:59:24.238537] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:06.988 [2024-12-05 18:59:24.370775] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:06.988 [2024-12-05 18:59:24.395263] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:06.988 [2024-12-05 18:59:24.438220] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:06.988 [2024-12-05 18:59:24.438255] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.559 [2024-12-05 18:59:25.061339] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:07.559 [2024-12-05 18:59:25.061407] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:07.559 [2024-12-05 18:59:25.061433] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:07.559 [2024-12-05 18:59:25.061444] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:07.559 [2024-12-05 18:59:25.061450] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:07.559 [2024-12-05 18:59:25.061461] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:07.559 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:07.559 "name": "Existed_Raid", 00:09:07.559 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:07.559 "strip_size_kb": 0, 00:09:07.559 "state": "configuring", 00:09:07.559 "raid_level": "raid1", 00:09:07.559 "superblock": false, 00:09:07.559 "num_base_bdevs": 3, 00:09:07.559 "num_base_bdevs_discovered": 0, 00:09:07.559 "num_base_bdevs_operational": 3, 00:09:07.559 "base_bdevs_list": [ 00:09:07.559 { 00:09:07.559 "name": "BaseBdev1", 00:09:07.559 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:07.559 "is_configured": false, 00:09:07.559 "data_offset": 0, 00:09:07.559 "data_size": 0 00:09:07.559 }, 00:09:07.559 { 00:09:07.559 "name": "BaseBdev2", 00:09:07.559 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:07.559 "is_configured": false, 00:09:07.559 "data_offset": 0, 00:09:07.559 "data_size": 0 00:09:07.559 }, 00:09:07.559 { 00:09:07.559 "name": "BaseBdev3", 00:09:07.559 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:07.559 "is_configured": false, 00:09:07.559 "data_offset": 0, 00:09:07.559 "data_size": 0 00:09:07.559 } 00:09:07.559 ] 00:09:07.559 }' 00:09:07.818 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:07.818 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.077 [2024-12-05 18:59:25.512474] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:08.077 [2024-12-05 18:59:25.512560] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.077 [2024-12-05 18:59:25.524480] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:08.077 [2024-12-05 18:59:25.524563] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:08.077 [2024-12-05 18:59:25.524589] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:08.077 [2024-12-05 18:59:25.524612] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:08.077 [2024-12-05 18:59:25.524629] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:08.077 [2024-12-05 18:59:25.524650] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.077 [2024-12-05 18:59:25.545448] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:08.077 BaseBdev1 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.077 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.077 [ 00:09:08.077 { 00:09:08.077 "name": "BaseBdev1", 00:09:08.077 "aliases": [ 00:09:08.077 "e07be9c9-f372-4d14-b222-80d5e7fa533f" 00:09:08.077 ], 00:09:08.077 "product_name": "Malloc disk", 00:09:08.077 "block_size": 512, 00:09:08.077 "num_blocks": 65536, 00:09:08.077 "uuid": "e07be9c9-f372-4d14-b222-80d5e7fa533f", 00:09:08.077 "assigned_rate_limits": { 00:09:08.077 "rw_ios_per_sec": 0, 00:09:08.077 "rw_mbytes_per_sec": 0, 00:09:08.077 "r_mbytes_per_sec": 0, 00:09:08.077 "w_mbytes_per_sec": 0 00:09:08.077 }, 00:09:08.077 "claimed": true, 00:09:08.077 "claim_type": "exclusive_write", 00:09:08.077 "zoned": false, 00:09:08.077 "supported_io_types": { 00:09:08.077 "read": true, 00:09:08.077 "write": true, 00:09:08.077 "unmap": true, 00:09:08.077 "flush": true, 00:09:08.077 "reset": true, 00:09:08.077 "nvme_admin": false, 00:09:08.077 "nvme_io": false, 00:09:08.077 "nvme_io_md": false, 00:09:08.077 "write_zeroes": true, 00:09:08.077 "zcopy": true, 00:09:08.077 "get_zone_info": false, 00:09:08.077 "zone_management": false, 00:09:08.077 "zone_append": false, 00:09:08.077 "compare": false, 00:09:08.077 "compare_and_write": false, 00:09:08.077 "abort": true, 00:09:08.077 "seek_hole": false, 00:09:08.077 "seek_data": false, 00:09:08.077 "copy": true, 00:09:08.077 "nvme_iov_md": false 00:09:08.077 }, 00:09:08.077 "memory_domains": [ 00:09:08.077 { 00:09:08.077 "dma_device_id": "system", 00:09:08.077 "dma_device_type": 1 00:09:08.077 }, 00:09:08.077 { 00:09:08.077 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:08.078 "dma_device_type": 2 00:09:08.078 } 00:09:08.078 ], 00:09:08.078 "driver_specific": {} 00:09:08.078 } 00:09:08.078 ] 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:08.078 "name": "Existed_Raid", 00:09:08.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.078 "strip_size_kb": 0, 00:09:08.078 "state": "configuring", 00:09:08.078 "raid_level": "raid1", 00:09:08.078 "superblock": false, 00:09:08.078 "num_base_bdevs": 3, 00:09:08.078 "num_base_bdevs_discovered": 1, 00:09:08.078 "num_base_bdevs_operational": 3, 00:09:08.078 "base_bdevs_list": [ 00:09:08.078 { 00:09:08.078 "name": "BaseBdev1", 00:09:08.078 "uuid": "e07be9c9-f372-4d14-b222-80d5e7fa533f", 00:09:08.078 "is_configured": true, 00:09:08.078 "data_offset": 0, 00:09:08.078 "data_size": 65536 00:09:08.078 }, 00:09:08.078 { 00:09:08.078 "name": "BaseBdev2", 00:09:08.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.078 "is_configured": false, 00:09:08.078 "data_offset": 0, 00:09:08.078 "data_size": 0 00:09:08.078 }, 00:09:08.078 { 00:09:08.078 "name": "BaseBdev3", 00:09:08.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.078 "is_configured": false, 00:09:08.078 "data_offset": 0, 00:09:08.078 "data_size": 0 00:09:08.078 } 00:09:08.078 ] 00:09:08.078 }' 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:08.078 18:59:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.646 [2024-12-05 18:59:26.008743] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:08.646 [2024-12-05 18:59:26.008782] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.646 [2024-12-05 18:59:26.016759] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:08.646 [2024-12-05 18:59:26.018578] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:08.646 [2024-12-05 18:59:26.018622] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:08.646 [2024-12-05 18:59:26.018632] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:08.646 [2024-12-05 18:59:26.018642] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.646 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:08.646 "name": "Existed_Raid", 00:09:08.646 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.646 "strip_size_kb": 0, 00:09:08.646 "state": "configuring", 00:09:08.646 "raid_level": "raid1", 00:09:08.646 "superblock": false, 00:09:08.646 "num_base_bdevs": 3, 00:09:08.646 "num_base_bdevs_discovered": 1, 00:09:08.646 "num_base_bdevs_operational": 3, 00:09:08.646 "base_bdevs_list": [ 00:09:08.646 { 00:09:08.646 "name": "BaseBdev1", 00:09:08.646 "uuid": "e07be9c9-f372-4d14-b222-80d5e7fa533f", 00:09:08.646 "is_configured": true, 00:09:08.646 "data_offset": 0, 00:09:08.646 "data_size": 65536 00:09:08.646 }, 00:09:08.646 { 00:09:08.646 "name": "BaseBdev2", 00:09:08.647 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.647 "is_configured": false, 00:09:08.647 "data_offset": 0, 00:09:08.647 "data_size": 0 00:09:08.647 }, 00:09:08.647 { 00:09:08.647 "name": "BaseBdev3", 00:09:08.647 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.647 "is_configured": false, 00:09:08.647 "data_offset": 0, 00:09:08.647 "data_size": 0 00:09:08.647 } 00:09:08.647 ] 00:09:08.647 }' 00:09:08.647 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:08.647 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.906 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:08.906 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.906 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.906 [2024-12-05 18:59:26.447094] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:08.906 BaseBdev2 00:09:08.906 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.906 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:08.906 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:08.906 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:08.906 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:08.906 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:08.906 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:08.906 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:08.906 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.907 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.907 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:08.907 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:08.907 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:08.907 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.165 [ 00:09:09.165 { 00:09:09.165 "name": "BaseBdev2", 00:09:09.165 "aliases": [ 00:09:09.165 "e7105857-418a-4d74-9e48-016dac1b20db" 00:09:09.165 ], 00:09:09.165 "product_name": "Malloc disk", 00:09:09.165 "block_size": 512, 00:09:09.165 "num_blocks": 65536, 00:09:09.165 "uuid": "e7105857-418a-4d74-9e48-016dac1b20db", 00:09:09.165 "assigned_rate_limits": { 00:09:09.165 "rw_ios_per_sec": 0, 00:09:09.165 "rw_mbytes_per_sec": 0, 00:09:09.165 "r_mbytes_per_sec": 0, 00:09:09.165 "w_mbytes_per_sec": 0 00:09:09.165 }, 00:09:09.165 "claimed": true, 00:09:09.165 "claim_type": "exclusive_write", 00:09:09.165 "zoned": false, 00:09:09.165 "supported_io_types": { 00:09:09.165 "read": true, 00:09:09.165 "write": true, 00:09:09.165 "unmap": true, 00:09:09.165 "flush": true, 00:09:09.165 "reset": true, 00:09:09.165 "nvme_admin": false, 00:09:09.165 "nvme_io": false, 00:09:09.165 "nvme_io_md": false, 00:09:09.165 "write_zeroes": true, 00:09:09.165 "zcopy": true, 00:09:09.165 "get_zone_info": false, 00:09:09.165 "zone_management": false, 00:09:09.165 "zone_append": false, 00:09:09.166 "compare": false, 00:09:09.166 "compare_and_write": false, 00:09:09.166 "abort": true, 00:09:09.166 "seek_hole": false, 00:09:09.166 "seek_data": false, 00:09:09.166 "copy": true, 00:09:09.166 "nvme_iov_md": false 00:09:09.166 }, 00:09:09.166 "memory_domains": [ 00:09:09.166 { 00:09:09.166 "dma_device_id": "system", 00:09:09.166 "dma_device_type": 1 00:09:09.166 }, 00:09:09.166 { 00:09:09.166 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.166 "dma_device_type": 2 00:09:09.166 } 00:09:09.166 ], 00:09:09.166 "driver_specific": {} 00:09:09.166 } 00:09:09.166 ] 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:09.166 "name": "Existed_Raid", 00:09:09.166 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:09.166 "strip_size_kb": 0, 00:09:09.166 "state": "configuring", 00:09:09.166 "raid_level": "raid1", 00:09:09.166 "superblock": false, 00:09:09.166 "num_base_bdevs": 3, 00:09:09.166 "num_base_bdevs_discovered": 2, 00:09:09.166 "num_base_bdevs_operational": 3, 00:09:09.166 "base_bdevs_list": [ 00:09:09.166 { 00:09:09.166 "name": "BaseBdev1", 00:09:09.166 "uuid": "e07be9c9-f372-4d14-b222-80d5e7fa533f", 00:09:09.166 "is_configured": true, 00:09:09.166 "data_offset": 0, 00:09:09.166 "data_size": 65536 00:09:09.166 }, 00:09:09.166 { 00:09:09.166 "name": "BaseBdev2", 00:09:09.166 "uuid": "e7105857-418a-4d74-9e48-016dac1b20db", 00:09:09.166 "is_configured": true, 00:09:09.166 "data_offset": 0, 00:09:09.166 "data_size": 65536 00:09:09.166 }, 00:09:09.166 { 00:09:09.166 "name": "BaseBdev3", 00:09:09.166 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:09.166 "is_configured": false, 00:09:09.166 "data_offset": 0, 00:09:09.166 "data_size": 0 00:09:09.166 } 00:09:09.166 ] 00:09:09.166 }' 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:09.166 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.425 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:09.425 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.425 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.425 [2024-12-05 18:59:26.943554] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:09.426 [2024-12-05 18:59:26.943897] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:09.426 [2024-12-05 18:59:26.944017] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:09:09.426 [2024-12-05 18:59:26.945154] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:09.426 [2024-12-05 18:59:26.945822] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:09.426 [2024-12-05 18:59:26.945880] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:09.426 BaseBdev3 00:09:09.426 [2024-12-05 18:59:26.946594] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:09.426 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.426 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:09.426 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:09.426 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:09.426 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:09.426 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:09.426 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:09.426 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:09.426 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.426 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.426 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.426 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:09.426 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.426 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.426 [ 00:09:09.426 { 00:09:09.426 "name": "BaseBdev3", 00:09:09.426 "aliases": [ 00:09:09.426 "543851b7-8132-459e-9c64-f81fd98e414d" 00:09:09.426 ], 00:09:09.426 "product_name": "Malloc disk", 00:09:09.426 "block_size": 512, 00:09:09.426 "num_blocks": 65536, 00:09:09.426 "uuid": "543851b7-8132-459e-9c64-f81fd98e414d", 00:09:09.426 "assigned_rate_limits": { 00:09:09.426 "rw_ios_per_sec": 0, 00:09:09.426 "rw_mbytes_per_sec": 0, 00:09:09.426 "r_mbytes_per_sec": 0, 00:09:09.426 "w_mbytes_per_sec": 0 00:09:09.426 }, 00:09:09.426 "claimed": true, 00:09:09.426 "claim_type": "exclusive_write", 00:09:09.426 "zoned": false, 00:09:09.426 "supported_io_types": { 00:09:09.426 "read": true, 00:09:09.426 "write": true, 00:09:09.426 "unmap": true, 00:09:09.426 "flush": true, 00:09:09.426 "reset": true, 00:09:09.426 "nvme_admin": false, 00:09:09.426 "nvme_io": false, 00:09:09.426 "nvme_io_md": false, 00:09:09.426 "write_zeroes": true, 00:09:09.426 "zcopy": true, 00:09:09.426 "get_zone_info": false, 00:09:09.426 "zone_management": false, 00:09:09.426 "zone_append": false, 00:09:09.426 "compare": false, 00:09:09.426 "compare_and_write": false, 00:09:09.426 "abort": true, 00:09:09.426 "seek_hole": false, 00:09:09.426 "seek_data": false, 00:09:09.426 "copy": true, 00:09:09.426 "nvme_iov_md": false 00:09:09.426 }, 00:09:09.426 "memory_domains": [ 00:09:09.426 { 00:09:09.426 "dma_device_id": "system", 00:09:09.426 "dma_device_type": 1 00:09:09.426 }, 00:09:09.686 { 00:09:09.686 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.686 "dma_device_type": 2 00:09:09.686 } 00:09:09.686 ], 00:09:09.686 "driver_specific": {} 00:09:09.686 } 00:09:09.686 ] 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.686 18:59:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.686 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.686 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:09.686 "name": "Existed_Raid", 00:09:09.686 "uuid": "4f4148e3-7ff2-4b8b-a95d-805d5d28ffde", 00:09:09.686 "strip_size_kb": 0, 00:09:09.686 "state": "online", 00:09:09.686 "raid_level": "raid1", 00:09:09.686 "superblock": false, 00:09:09.686 "num_base_bdevs": 3, 00:09:09.686 "num_base_bdevs_discovered": 3, 00:09:09.686 "num_base_bdevs_operational": 3, 00:09:09.686 "base_bdevs_list": [ 00:09:09.686 { 00:09:09.686 "name": "BaseBdev1", 00:09:09.686 "uuid": "e07be9c9-f372-4d14-b222-80d5e7fa533f", 00:09:09.686 "is_configured": true, 00:09:09.686 "data_offset": 0, 00:09:09.686 "data_size": 65536 00:09:09.686 }, 00:09:09.686 { 00:09:09.686 "name": "BaseBdev2", 00:09:09.686 "uuid": "e7105857-418a-4d74-9e48-016dac1b20db", 00:09:09.686 "is_configured": true, 00:09:09.686 "data_offset": 0, 00:09:09.686 "data_size": 65536 00:09:09.686 }, 00:09:09.686 { 00:09:09.686 "name": "BaseBdev3", 00:09:09.686 "uuid": "543851b7-8132-459e-9c64-f81fd98e414d", 00:09:09.686 "is_configured": true, 00:09:09.686 "data_offset": 0, 00:09:09.686 "data_size": 65536 00:09:09.686 } 00:09:09.686 ] 00:09:09.686 }' 00:09:09.686 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:09.686 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.946 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:09.946 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:09.946 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:09.946 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:09.946 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:09.946 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:09.946 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:09.946 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:09.946 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.946 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:09.946 [2024-12-05 18:59:27.411007] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:09.946 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:09.946 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:09.946 "name": "Existed_Raid", 00:09:09.946 "aliases": [ 00:09:09.946 "4f4148e3-7ff2-4b8b-a95d-805d5d28ffde" 00:09:09.946 ], 00:09:09.946 "product_name": "Raid Volume", 00:09:09.946 "block_size": 512, 00:09:09.946 "num_blocks": 65536, 00:09:09.946 "uuid": "4f4148e3-7ff2-4b8b-a95d-805d5d28ffde", 00:09:09.946 "assigned_rate_limits": { 00:09:09.946 "rw_ios_per_sec": 0, 00:09:09.946 "rw_mbytes_per_sec": 0, 00:09:09.946 "r_mbytes_per_sec": 0, 00:09:09.946 "w_mbytes_per_sec": 0 00:09:09.946 }, 00:09:09.946 "claimed": false, 00:09:09.946 "zoned": false, 00:09:09.946 "supported_io_types": { 00:09:09.946 "read": true, 00:09:09.946 "write": true, 00:09:09.946 "unmap": false, 00:09:09.946 "flush": false, 00:09:09.946 "reset": true, 00:09:09.946 "nvme_admin": false, 00:09:09.946 "nvme_io": false, 00:09:09.946 "nvme_io_md": false, 00:09:09.946 "write_zeroes": true, 00:09:09.946 "zcopy": false, 00:09:09.946 "get_zone_info": false, 00:09:09.946 "zone_management": false, 00:09:09.946 "zone_append": false, 00:09:09.946 "compare": false, 00:09:09.946 "compare_and_write": false, 00:09:09.946 "abort": false, 00:09:09.946 "seek_hole": false, 00:09:09.946 "seek_data": false, 00:09:09.946 "copy": false, 00:09:09.946 "nvme_iov_md": false 00:09:09.946 }, 00:09:09.946 "memory_domains": [ 00:09:09.946 { 00:09:09.946 "dma_device_id": "system", 00:09:09.946 "dma_device_type": 1 00:09:09.946 }, 00:09:09.946 { 00:09:09.946 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.946 "dma_device_type": 2 00:09:09.946 }, 00:09:09.946 { 00:09:09.946 "dma_device_id": "system", 00:09:09.946 "dma_device_type": 1 00:09:09.946 }, 00:09:09.946 { 00:09:09.946 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.946 "dma_device_type": 2 00:09:09.946 }, 00:09:09.946 { 00:09:09.946 "dma_device_id": "system", 00:09:09.946 "dma_device_type": 1 00:09:09.946 }, 00:09:09.946 { 00:09:09.946 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.946 "dma_device_type": 2 00:09:09.946 } 00:09:09.946 ], 00:09:09.946 "driver_specific": { 00:09:09.946 "raid": { 00:09:09.946 "uuid": "4f4148e3-7ff2-4b8b-a95d-805d5d28ffde", 00:09:09.946 "strip_size_kb": 0, 00:09:09.946 "state": "online", 00:09:09.946 "raid_level": "raid1", 00:09:09.946 "superblock": false, 00:09:09.946 "num_base_bdevs": 3, 00:09:09.946 "num_base_bdevs_discovered": 3, 00:09:09.946 "num_base_bdevs_operational": 3, 00:09:09.946 "base_bdevs_list": [ 00:09:09.946 { 00:09:09.946 "name": "BaseBdev1", 00:09:09.946 "uuid": "e07be9c9-f372-4d14-b222-80d5e7fa533f", 00:09:09.946 "is_configured": true, 00:09:09.946 "data_offset": 0, 00:09:09.946 "data_size": 65536 00:09:09.946 }, 00:09:09.946 { 00:09:09.946 "name": "BaseBdev2", 00:09:09.946 "uuid": "e7105857-418a-4d74-9e48-016dac1b20db", 00:09:09.946 "is_configured": true, 00:09:09.946 "data_offset": 0, 00:09:09.946 "data_size": 65536 00:09:09.946 }, 00:09:09.946 { 00:09:09.946 "name": "BaseBdev3", 00:09:09.946 "uuid": "543851b7-8132-459e-9c64-f81fd98e414d", 00:09:09.946 "is_configured": true, 00:09:09.946 "data_offset": 0, 00:09:09.946 "data_size": 65536 00:09:09.946 } 00:09:09.946 ] 00:09:09.946 } 00:09:09.946 } 00:09:09.946 }' 00:09:09.946 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:09.946 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:09.946 BaseBdev2 00:09:09.946 BaseBdev3' 00:09:09.946 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.207 [2024-12-05 18:59:27.658312] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.207 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:10.207 "name": "Existed_Raid", 00:09:10.207 "uuid": "4f4148e3-7ff2-4b8b-a95d-805d5d28ffde", 00:09:10.207 "strip_size_kb": 0, 00:09:10.207 "state": "online", 00:09:10.207 "raid_level": "raid1", 00:09:10.207 "superblock": false, 00:09:10.207 "num_base_bdevs": 3, 00:09:10.207 "num_base_bdevs_discovered": 2, 00:09:10.208 "num_base_bdevs_operational": 2, 00:09:10.208 "base_bdevs_list": [ 00:09:10.208 { 00:09:10.208 "name": null, 00:09:10.208 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:10.208 "is_configured": false, 00:09:10.208 "data_offset": 0, 00:09:10.208 "data_size": 65536 00:09:10.208 }, 00:09:10.208 { 00:09:10.208 "name": "BaseBdev2", 00:09:10.208 "uuid": "e7105857-418a-4d74-9e48-016dac1b20db", 00:09:10.208 "is_configured": true, 00:09:10.208 "data_offset": 0, 00:09:10.208 "data_size": 65536 00:09:10.208 }, 00:09:10.208 { 00:09:10.208 "name": "BaseBdev3", 00:09:10.208 "uuid": "543851b7-8132-459e-9c64-f81fd98e414d", 00:09:10.208 "is_configured": true, 00:09:10.208 "data_offset": 0, 00:09:10.208 "data_size": 65536 00:09:10.208 } 00:09:10.208 ] 00:09:10.208 }' 00:09:10.208 18:59:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:10.208 18:59:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.779 [2024-12-05 18:59:28.188784] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.779 [2024-12-05 18:59:28.259792] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:10.779 [2024-12-05 18:59:28.259875] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:10.779 [2024-12-05 18:59:28.271530] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:10.779 [2024-12-05 18:59:28.271644] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:10.779 [2024-12-05 18:59:28.271732] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.779 BaseBdev2 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:10.779 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.779 [ 00:09:10.779 { 00:09:10.779 "name": "BaseBdev2", 00:09:10.779 "aliases": [ 00:09:10.779 "0c7143c6-fbe9-4c68-9a03-8c2831e5335a" 00:09:10.779 ], 00:09:10.779 "product_name": "Malloc disk", 00:09:10.779 "block_size": 512, 00:09:10.779 "num_blocks": 65536, 00:09:10.779 "uuid": "0c7143c6-fbe9-4c68-9a03-8c2831e5335a", 00:09:10.779 "assigned_rate_limits": { 00:09:10.779 "rw_ios_per_sec": 0, 00:09:10.779 "rw_mbytes_per_sec": 0, 00:09:10.779 "r_mbytes_per_sec": 0, 00:09:10.779 "w_mbytes_per_sec": 0 00:09:10.779 }, 00:09:10.779 "claimed": false, 00:09:11.040 "zoned": false, 00:09:11.040 "supported_io_types": { 00:09:11.040 "read": true, 00:09:11.040 "write": true, 00:09:11.040 "unmap": true, 00:09:11.040 "flush": true, 00:09:11.040 "reset": true, 00:09:11.040 "nvme_admin": false, 00:09:11.040 "nvme_io": false, 00:09:11.040 "nvme_io_md": false, 00:09:11.040 "write_zeroes": true, 00:09:11.040 "zcopy": true, 00:09:11.040 "get_zone_info": false, 00:09:11.040 "zone_management": false, 00:09:11.040 "zone_append": false, 00:09:11.040 "compare": false, 00:09:11.040 "compare_and_write": false, 00:09:11.040 "abort": true, 00:09:11.040 "seek_hole": false, 00:09:11.040 "seek_data": false, 00:09:11.040 "copy": true, 00:09:11.040 "nvme_iov_md": false 00:09:11.040 }, 00:09:11.040 "memory_domains": [ 00:09:11.040 { 00:09:11.040 "dma_device_id": "system", 00:09:11.040 "dma_device_type": 1 00:09:11.040 }, 00:09:11.040 { 00:09:11.040 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:11.040 "dma_device_type": 2 00:09:11.040 } 00:09:11.040 ], 00:09:11.040 "driver_specific": {} 00:09:11.040 } 00:09:11.040 ] 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.040 BaseBdev3 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.040 [ 00:09:11.040 { 00:09:11.040 "name": "BaseBdev3", 00:09:11.040 "aliases": [ 00:09:11.040 "3a9adcf9-26a3-450e-ae1a-899a95a5a01d" 00:09:11.040 ], 00:09:11.040 "product_name": "Malloc disk", 00:09:11.040 "block_size": 512, 00:09:11.040 "num_blocks": 65536, 00:09:11.040 "uuid": "3a9adcf9-26a3-450e-ae1a-899a95a5a01d", 00:09:11.040 "assigned_rate_limits": { 00:09:11.040 "rw_ios_per_sec": 0, 00:09:11.040 "rw_mbytes_per_sec": 0, 00:09:11.040 "r_mbytes_per_sec": 0, 00:09:11.040 "w_mbytes_per_sec": 0 00:09:11.040 }, 00:09:11.040 "claimed": false, 00:09:11.040 "zoned": false, 00:09:11.040 "supported_io_types": { 00:09:11.040 "read": true, 00:09:11.040 "write": true, 00:09:11.040 "unmap": true, 00:09:11.040 "flush": true, 00:09:11.040 "reset": true, 00:09:11.040 "nvme_admin": false, 00:09:11.040 "nvme_io": false, 00:09:11.040 "nvme_io_md": false, 00:09:11.040 "write_zeroes": true, 00:09:11.040 "zcopy": true, 00:09:11.040 "get_zone_info": false, 00:09:11.040 "zone_management": false, 00:09:11.040 "zone_append": false, 00:09:11.040 "compare": false, 00:09:11.040 "compare_and_write": false, 00:09:11.040 "abort": true, 00:09:11.040 "seek_hole": false, 00:09:11.040 "seek_data": false, 00:09:11.040 "copy": true, 00:09:11.040 "nvme_iov_md": false 00:09:11.040 }, 00:09:11.040 "memory_domains": [ 00:09:11.040 { 00:09:11.040 "dma_device_id": "system", 00:09:11.040 "dma_device_type": 1 00:09:11.040 }, 00:09:11.040 { 00:09:11.040 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:11.040 "dma_device_type": 2 00:09:11.040 } 00:09:11.040 ], 00:09:11.040 "driver_specific": {} 00:09:11.040 } 00:09:11.040 ] 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.040 [2024-12-05 18:59:28.375803] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:11.040 [2024-12-05 18:59:28.375847] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:11.040 [2024-12-05 18:59:28.375870] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:11.040 [2024-12-05 18:59:28.377697] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.040 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.041 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:11.041 "name": "Existed_Raid", 00:09:11.041 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:11.041 "strip_size_kb": 0, 00:09:11.041 "state": "configuring", 00:09:11.041 "raid_level": "raid1", 00:09:11.041 "superblock": false, 00:09:11.041 "num_base_bdevs": 3, 00:09:11.041 "num_base_bdevs_discovered": 2, 00:09:11.041 "num_base_bdevs_operational": 3, 00:09:11.041 "base_bdevs_list": [ 00:09:11.041 { 00:09:11.041 "name": "BaseBdev1", 00:09:11.041 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:11.041 "is_configured": false, 00:09:11.041 "data_offset": 0, 00:09:11.041 "data_size": 0 00:09:11.041 }, 00:09:11.041 { 00:09:11.041 "name": "BaseBdev2", 00:09:11.041 "uuid": "0c7143c6-fbe9-4c68-9a03-8c2831e5335a", 00:09:11.041 "is_configured": true, 00:09:11.041 "data_offset": 0, 00:09:11.041 "data_size": 65536 00:09:11.041 }, 00:09:11.041 { 00:09:11.041 "name": "BaseBdev3", 00:09:11.041 "uuid": "3a9adcf9-26a3-450e-ae1a-899a95a5a01d", 00:09:11.041 "is_configured": true, 00:09:11.041 "data_offset": 0, 00:09:11.041 "data_size": 65536 00:09:11.041 } 00:09:11.041 ] 00:09:11.041 }' 00:09:11.041 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:11.041 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.301 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:11.301 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.301 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.301 [2024-12-05 18:59:28.795070] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:11.301 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.301 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:11.301 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:11.301 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:11.301 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:11.302 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:11.302 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:11.302 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.302 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.302 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.302 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.302 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.302 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.302 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.302 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:11.302 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.302 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:11.302 "name": "Existed_Raid", 00:09:11.302 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:11.302 "strip_size_kb": 0, 00:09:11.302 "state": "configuring", 00:09:11.302 "raid_level": "raid1", 00:09:11.302 "superblock": false, 00:09:11.302 "num_base_bdevs": 3, 00:09:11.302 "num_base_bdevs_discovered": 1, 00:09:11.302 "num_base_bdevs_operational": 3, 00:09:11.302 "base_bdevs_list": [ 00:09:11.302 { 00:09:11.302 "name": "BaseBdev1", 00:09:11.302 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:11.302 "is_configured": false, 00:09:11.302 "data_offset": 0, 00:09:11.302 "data_size": 0 00:09:11.302 }, 00:09:11.302 { 00:09:11.302 "name": null, 00:09:11.302 "uuid": "0c7143c6-fbe9-4c68-9a03-8c2831e5335a", 00:09:11.302 "is_configured": false, 00:09:11.302 "data_offset": 0, 00:09:11.302 "data_size": 65536 00:09:11.302 }, 00:09:11.302 { 00:09:11.302 "name": "BaseBdev3", 00:09:11.302 "uuid": "3a9adcf9-26a3-450e-ae1a-899a95a5a01d", 00:09:11.302 "is_configured": true, 00:09:11.302 "data_offset": 0, 00:09:11.302 "data_size": 65536 00:09:11.302 } 00:09:11.302 ] 00:09:11.302 }' 00:09:11.302 18:59:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:11.302 18:59:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.872 [2024-12-05 18:59:29.213542] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:11.872 BaseBdev1 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.872 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.873 [ 00:09:11.873 { 00:09:11.873 "name": "BaseBdev1", 00:09:11.873 "aliases": [ 00:09:11.873 "74696c10-5c7e-44c9-9a83-b576d8e7df73" 00:09:11.873 ], 00:09:11.873 "product_name": "Malloc disk", 00:09:11.873 "block_size": 512, 00:09:11.873 "num_blocks": 65536, 00:09:11.873 "uuid": "74696c10-5c7e-44c9-9a83-b576d8e7df73", 00:09:11.873 "assigned_rate_limits": { 00:09:11.873 "rw_ios_per_sec": 0, 00:09:11.873 "rw_mbytes_per_sec": 0, 00:09:11.873 "r_mbytes_per_sec": 0, 00:09:11.873 "w_mbytes_per_sec": 0 00:09:11.873 }, 00:09:11.873 "claimed": true, 00:09:11.873 "claim_type": "exclusive_write", 00:09:11.873 "zoned": false, 00:09:11.873 "supported_io_types": { 00:09:11.873 "read": true, 00:09:11.873 "write": true, 00:09:11.873 "unmap": true, 00:09:11.873 "flush": true, 00:09:11.873 "reset": true, 00:09:11.873 "nvme_admin": false, 00:09:11.873 "nvme_io": false, 00:09:11.873 "nvme_io_md": false, 00:09:11.873 "write_zeroes": true, 00:09:11.873 "zcopy": true, 00:09:11.873 "get_zone_info": false, 00:09:11.873 "zone_management": false, 00:09:11.873 "zone_append": false, 00:09:11.873 "compare": false, 00:09:11.873 "compare_and_write": false, 00:09:11.873 "abort": true, 00:09:11.873 "seek_hole": false, 00:09:11.873 "seek_data": false, 00:09:11.873 "copy": true, 00:09:11.873 "nvme_iov_md": false 00:09:11.873 }, 00:09:11.873 "memory_domains": [ 00:09:11.873 { 00:09:11.873 "dma_device_id": "system", 00:09:11.873 "dma_device_type": 1 00:09:11.873 }, 00:09:11.873 { 00:09:11.873 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:11.873 "dma_device_type": 2 00:09:11.873 } 00:09:11.873 ], 00:09:11.873 "driver_specific": {} 00:09:11.873 } 00:09:11.873 ] 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:11.873 "name": "Existed_Raid", 00:09:11.873 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:11.873 "strip_size_kb": 0, 00:09:11.873 "state": "configuring", 00:09:11.873 "raid_level": "raid1", 00:09:11.873 "superblock": false, 00:09:11.873 "num_base_bdevs": 3, 00:09:11.873 "num_base_bdevs_discovered": 2, 00:09:11.873 "num_base_bdevs_operational": 3, 00:09:11.873 "base_bdevs_list": [ 00:09:11.873 { 00:09:11.873 "name": "BaseBdev1", 00:09:11.873 "uuid": "74696c10-5c7e-44c9-9a83-b576d8e7df73", 00:09:11.873 "is_configured": true, 00:09:11.873 "data_offset": 0, 00:09:11.873 "data_size": 65536 00:09:11.873 }, 00:09:11.873 { 00:09:11.873 "name": null, 00:09:11.873 "uuid": "0c7143c6-fbe9-4c68-9a03-8c2831e5335a", 00:09:11.873 "is_configured": false, 00:09:11.873 "data_offset": 0, 00:09:11.873 "data_size": 65536 00:09:11.873 }, 00:09:11.873 { 00:09:11.873 "name": "BaseBdev3", 00:09:11.873 "uuid": "3a9adcf9-26a3-450e-ae1a-899a95a5a01d", 00:09:11.873 "is_configured": true, 00:09:11.873 "data_offset": 0, 00:09:11.873 "data_size": 65536 00:09:11.873 } 00:09:11.873 ] 00:09:11.873 }' 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:11.873 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.145 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:12.145 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.145 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.145 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.145 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.145 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:12.145 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:12.145 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.145 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.145 [2024-12-05 18:59:29.700767] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:12.405 "name": "Existed_Raid", 00:09:12.405 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:12.405 "strip_size_kb": 0, 00:09:12.405 "state": "configuring", 00:09:12.405 "raid_level": "raid1", 00:09:12.405 "superblock": false, 00:09:12.405 "num_base_bdevs": 3, 00:09:12.405 "num_base_bdevs_discovered": 1, 00:09:12.405 "num_base_bdevs_operational": 3, 00:09:12.405 "base_bdevs_list": [ 00:09:12.405 { 00:09:12.405 "name": "BaseBdev1", 00:09:12.405 "uuid": "74696c10-5c7e-44c9-9a83-b576d8e7df73", 00:09:12.405 "is_configured": true, 00:09:12.405 "data_offset": 0, 00:09:12.405 "data_size": 65536 00:09:12.405 }, 00:09:12.405 { 00:09:12.405 "name": null, 00:09:12.405 "uuid": "0c7143c6-fbe9-4c68-9a03-8c2831e5335a", 00:09:12.405 "is_configured": false, 00:09:12.405 "data_offset": 0, 00:09:12.405 "data_size": 65536 00:09:12.405 }, 00:09:12.405 { 00:09:12.405 "name": null, 00:09:12.405 "uuid": "3a9adcf9-26a3-450e-ae1a-899a95a5a01d", 00:09:12.405 "is_configured": false, 00:09:12.405 "data_offset": 0, 00:09:12.405 "data_size": 65536 00:09:12.405 } 00:09:12.405 ] 00:09:12.405 }' 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:12.405 18:59:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.664 [2024-12-05 18:59:30.148008] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:12.664 "name": "Existed_Raid", 00:09:12.664 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:12.664 "strip_size_kb": 0, 00:09:12.664 "state": "configuring", 00:09:12.664 "raid_level": "raid1", 00:09:12.664 "superblock": false, 00:09:12.664 "num_base_bdevs": 3, 00:09:12.664 "num_base_bdevs_discovered": 2, 00:09:12.664 "num_base_bdevs_operational": 3, 00:09:12.664 "base_bdevs_list": [ 00:09:12.664 { 00:09:12.664 "name": "BaseBdev1", 00:09:12.664 "uuid": "74696c10-5c7e-44c9-9a83-b576d8e7df73", 00:09:12.664 "is_configured": true, 00:09:12.664 "data_offset": 0, 00:09:12.664 "data_size": 65536 00:09:12.664 }, 00:09:12.664 { 00:09:12.664 "name": null, 00:09:12.664 "uuid": "0c7143c6-fbe9-4c68-9a03-8c2831e5335a", 00:09:12.664 "is_configured": false, 00:09:12.664 "data_offset": 0, 00:09:12.664 "data_size": 65536 00:09:12.664 }, 00:09:12.664 { 00:09:12.664 "name": "BaseBdev3", 00:09:12.664 "uuid": "3a9adcf9-26a3-450e-ae1a-899a95a5a01d", 00:09:12.664 "is_configured": true, 00:09:12.664 "data_offset": 0, 00:09:12.664 "data_size": 65536 00:09:12.664 } 00:09:12.664 ] 00:09:12.664 }' 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:12.664 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.233 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:13.233 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:13.233 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.233 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.234 [2024-12-05 18:59:30.591285] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:13.234 "name": "Existed_Raid", 00:09:13.234 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:13.234 "strip_size_kb": 0, 00:09:13.234 "state": "configuring", 00:09:13.234 "raid_level": "raid1", 00:09:13.234 "superblock": false, 00:09:13.234 "num_base_bdevs": 3, 00:09:13.234 "num_base_bdevs_discovered": 1, 00:09:13.234 "num_base_bdevs_operational": 3, 00:09:13.234 "base_bdevs_list": [ 00:09:13.234 { 00:09:13.234 "name": null, 00:09:13.234 "uuid": "74696c10-5c7e-44c9-9a83-b576d8e7df73", 00:09:13.234 "is_configured": false, 00:09:13.234 "data_offset": 0, 00:09:13.234 "data_size": 65536 00:09:13.234 }, 00:09:13.234 { 00:09:13.234 "name": null, 00:09:13.234 "uuid": "0c7143c6-fbe9-4c68-9a03-8c2831e5335a", 00:09:13.234 "is_configured": false, 00:09:13.234 "data_offset": 0, 00:09:13.234 "data_size": 65536 00:09:13.234 }, 00:09:13.234 { 00:09:13.234 "name": "BaseBdev3", 00:09:13.234 "uuid": "3a9adcf9-26a3-450e-ae1a-899a95a5a01d", 00:09:13.234 "is_configured": true, 00:09:13.234 "data_offset": 0, 00:09:13.234 "data_size": 65536 00:09:13.234 } 00:09:13.234 ] 00:09:13.234 }' 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:13.234 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.494 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:13.495 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:13.495 18:59:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.495 18:59:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.495 [2024-12-05 18:59:31.041095] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:13.495 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.755 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:13.755 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:13.755 "name": "Existed_Raid", 00:09:13.755 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:13.755 "strip_size_kb": 0, 00:09:13.755 "state": "configuring", 00:09:13.755 "raid_level": "raid1", 00:09:13.755 "superblock": false, 00:09:13.755 "num_base_bdevs": 3, 00:09:13.755 "num_base_bdevs_discovered": 2, 00:09:13.755 "num_base_bdevs_operational": 3, 00:09:13.755 "base_bdevs_list": [ 00:09:13.755 { 00:09:13.755 "name": null, 00:09:13.755 "uuid": "74696c10-5c7e-44c9-9a83-b576d8e7df73", 00:09:13.755 "is_configured": false, 00:09:13.755 "data_offset": 0, 00:09:13.755 "data_size": 65536 00:09:13.755 }, 00:09:13.755 { 00:09:13.755 "name": "BaseBdev2", 00:09:13.755 "uuid": "0c7143c6-fbe9-4c68-9a03-8c2831e5335a", 00:09:13.755 "is_configured": true, 00:09:13.755 "data_offset": 0, 00:09:13.755 "data_size": 65536 00:09:13.755 }, 00:09:13.755 { 00:09:13.755 "name": "BaseBdev3", 00:09:13.755 "uuid": "3a9adcf9-26a3-450e-ae1a-899a95a5a01d", 00:09:13.755 "is_configured": true, 00:09:13.755 "data_offset": 0, 00:09:13.755 "data_size": 65536 00:09:13.755 } 00:09:13.755 ] 00:09:13.755 }' 00:09:13.755 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:13.755 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 74696c10-5c7e-44c9-9a83-b576d8e7df73 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.016 [2024-12-05 18:59:31.558854] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:14.016 [2024-12-05 18:59:31.558897] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:14.016 [2024-12-05 18:59:31.558905] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:09:14.016 [2024-12-05 18:59:31.559149] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:09:14.016 [2024-12-05 18:59:31.559259] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:14.016 [2024-12-05 18:59:31.559272] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:14.016 [2024-12-05 18:59:31.559440] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:14.016 NewBaseBdev 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.016 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.276 [ 00:09:14.276 { 00:09:14.276 "name": "NewBaseBdev", 00:09:14.276 "aliases": [ 00:09:14.276 "74696c10-5c7e-44c9-9a83-b576d8e7df73" 00:09:14.276 ], 00:09:14.276 "product_name": "Malloc disk", 00:09:14.276 "block_size": 512, 00:09:14.276 "num_blocks": 65536, 00:09:14.276 "uuid": "74696c10-5c7e-44c9-9a83-b576d8e7df73", 00:09:14.276 "assigned_rate_limits": { 00:09:14.276 "rw_ios_per_sec": 0, 00:09:14.276 "rw_mbytes_per_sec": 0, 00:09:14.276 "r_mbytes_per_sec": 0, 00:09:14.277 "w_mbytes_per_sec": 0 00:09:14.277 }, 00:09:14.277 "claimed": true, 00:09:14.277 "claim_type": "exclusive_write", 00:09:14.277 "zoned": false, 00:09:14.277 "supported_io_types": { 00:09:14.277 "read": true, 00:09:14.277 "write": true, 00:09:14.277 "unmap": true, 00:09:14.277 "flush": true, 00:09:14.277 "reset": true, 00:09:14.277 "nvme_admin": false, 00:09:14.277 "nvme_io": false, 00:09:14.277 "nvme_io_md": false, 00:09:14.277 "write_zeroes": true, 00:09:14.277 "zcopy": true, 00:09:14.277 "get_zone_info": false, 00:09:14.277 "zone_management": false, 00:09:14.277 "zone_append": false, 00:09:14.277 "compare": false, 00:09:14.277 "compare_and_write": false, 00:09:14.277 "abort": true, 00:09:14.277 "seek_hole": false, 00:09:14.277 "seek_data": false, 00:09:14.277 "copy": true, 00:09:14.277 "nvme_iov_md": false 00:09:14.277 }, 00:09:14.277 "memory_domains": [ 00:09:14.277 { 00:09:14.277 "dma_device_id": "system", 00:09:14.277 "dma_device_type": 1 00:09:14.277 }, 00:09:14.277 { 00:09:14.277 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:14.277 "dma_device_type": 2 00:09:14.277 } 00:09:14.277 ], 00:09:14.277 "driver_specific": {} 00:09:14.277 } 00:09:14.277 ] 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:14.277 "name": "Existed_Raid", 00:09:14.277 "uuid": "491b6642-fb30-410e-b103-ba6d49b40865", 00:09:14.277 "strip_size_kb": 0, 00:09:14.277 "state": "online", 00:09:14.277 "raid_level": "raid1", 00:09:14.277 "superblock": false, 00:09:14.277 "num_base_bdevs": 3, 00:09:14.277 "num_base_bdevs_discovered": 3, 00:09:14.277 "num_base_bdevs_operational": 3, 00:09:14.277 "base_bdevs_list": [ 00:09:14.277 { 00:09:14.277 "name": "NewBaseBdev", 00:09:14.277 "uuid": "74696c10-5c7e-44c9-9a83-b576d8e7df73", 00:09:14.277 "is_configured": true, 00:09:14.277 "data_offset": 0, 00:09:14.277 "data_size": 65536 00:09:14.277 }, 00:09:14.277 { 00:09:14.277 "name": "BaseBdev2", 00:09:14.277 "uuid": "0c7143c6-fbe9-4c68-9a03-8c2831e5335a", 00:09:14.277 "is_configured": true, 00:09:14.277 "data_offset": 0, 00:09:14.277 "data_size": 65536 00:09:14.277 }, 00:09:14.277 { 00:09:14.277 "name": "BaseBdev3", 00:09:14.277 "uuid": "3a9adcf9-26a3-450e-ae1a-899a95a5a01d", 00:09:14.277 "is_configured": true, 00:09:14.277 "data_offset": 0, 00:09:14.277 "data_size": 65536 00:09:14.277 } 00:09:14.277 ] 00:09:14.277 }' 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:14.277 18:59:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.538 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:14.538 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:14.538 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:14.538 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:14.538 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:14.538 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:14.538 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:14.538 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:14.538 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.538 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.538 [2024-12-05 18:59:32.018382] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:14.538 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.538 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:14.538 "name": "Existed_Raid", 00:09:14.538 "aliases": [ 00:09:14.538 "491b6642-fb30-410e-b103-ba6d49b40865" 00:09:14.538 ], 00:09:14.538 "product_name": "Raid Volume", 00:09:14.538 "block_size": 512, 00:09:14.538 "num_blocks": 65536, 00:09:14.538 "uuid": "491b6642-fb30-410e-b103-ba6d49b40865", 00:09:14.538 "assigned_rate_limits": { 00:09:14.538 "rw_ios_per_sec": 0, 00:09:14.538 "rw_mbytes_per_sec": 0, 00:09:14.538 "r_mbytes_per_sec": 0, 00:09:14.538 "w_mbytes_per_sec": 0 00:09:14.538 }, 00:09:14.538 "claimed": false, 00:09:14.538 "zoned": false, 00:09:14.538 "supported_io_types": { 00:09:14.538 "read": true, 00:09:14.538 "write": true, 00:09:14.538 "unmap": false, 00:09:14.538 "flush": false, 00:09:14.538 "reset": true, 00:09:14.538 "nvme_admin": false, 00:09:14.538 "nvme_io": false, 00:09:14.538 "nvme_io_md": false, 00:09:14.538 "write_zeroes": true, 00:09:14.538 "zcopy": false, 00:09:14.538 "get_zone_info": false, 00:09:14.538 "zone_management": false, 00:09:14.538 "zone_append": false, 00:09:14.538 "compare": false, 00:09:14.538 "compare_and_write": false, 00:09:14.538 "abort": false, 00:09:14.538 "seek_hole": false, 00:09:14.538 "seek_data": false, 00:09:14.538 "copy": false, 00:09:14.538 "nvme_iov_md": false 00:09:14.538 }, 00:09:14.538 "memory_domains": [ 00:09:14.538 { 00:09:14.538 "dma_device_id": "system", 00:09:14.538 "dma_device_type": 1 00:09:14.538 }, 00:09:14.538 { 00:09:14.538 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:14.538 "dma_device_type": 2 00:09:14.538 }, 00:09:14.538 { 00:09:14.538 "dma_device_id": "system", 00:09:14.538 "dma_device_type": 1 00:09:14.538 }, 00:09:14.538 { 00:09:14.538 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:14.538 "dma_device_type": 2 00:09:14.538 }, 00:09:14.538 { 00:09:14.538 "dma_device_id": "system", 00:09:14.538 "dma_device_type": 1 00:09:14.538 }, 00:09:14.538 { 00:09:14.538 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:14.538 "dma_device_type": 2 00:09:14.538 } 00:09:14.538 ], 00:09:14.538 "driver_specific": { 00:09:14.538 "raid": { 00:09:14.538 "uuid": "491b6642-fb30-410e-b103-ba6d49b40865", 00:09:14.538 "strip_size_kb": 0, 00:09:14.538 "state": "online", 00:09:14.538 "raid_level": "raid1", 00:09:14.538 "superblock": false, 00:09:14.538 "num_base_bdevs": 3, 00:09:14.538 "num_base_bdevs_discovered": 3, 00:09:14.538 "num_base_bdevs_operational": 3, 00:09:14.538 "base_bdevs_list": [ 00:09:14.538 { 00:09:14.538 "name": "NewBaseBdev", 00:09:14.538 "uuid": "74696c10-5c7e-44c9-9a83-b576d8e7df73", 00:09:14.538 "is_configured": true, 00:09:14.538 "data_offset": 0, 00:09:14.538 "data_size": 65536 00:09:14.538 }, 00:09:14.538 { 00:09:14.538 "name": "BaseBdev2", 00:09:14.538 "uuid": "0c7143c6-fbe9-4c68-9a03-8c2831e5335a", 00:09:14.538 "is_configured": true, 00:09:14.538 "data_offset": 0, 00:09:14.538 "data_size": 65536 00:09:14.538 }, 00:09:14.538 { 00:09:14.538 "name": "BaseBdev3", 00:09:14.538 "uuid": "3a9adcf9-26a3-450e-ae1a-899a95a5a01d", 00:09:14.538 "is_configured": true, 00:09:14.538 "data_offset": 0, 00:09:14.538 "data_size": 65536 00:09:14.538 } 00:09:14.538 ] 00:09:14.538 } 00:09:14.538 } 00:09:14.538 }' 00:09:14.538 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:14.538 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:14.538 BaseBdev2 00:09:14.538 BaseBdev3' 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.799 [2024-12-05 18:59:32.293622] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:14.799 [2024-12-05 18:59:32.293649] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:14.799 [2024-12-05 18:59:32.293732] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:14.799 [2024-12-05 18:59:32.293988] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:14.799 [2024-12-05 18:59:32.294003] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 77965 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 77965 ']' 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 77965 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 77965 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 77965' 00:09:14.799 killing process with pid 77965 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 77965 00:09:14.799 [2024-12-05 18:59:32.332578] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:14.799 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 77965 00:09:15.059 [2024-12-05 18:59:32.362662] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:15.059 18:59:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:15.059 00:09:15.059 real 0m8.422s 00:09:15.059 user 0m14.488s 00:09:15.059 sys 0m1.556s 00:09:15.059 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:15.059 18:59:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.059 ************************************ 00:09:15.059 END TEST raid_state_function_test 00:09:15.059 ************************************ 00:09:15.328 18:59:32 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 3 true 00:09:15.328 18:59:32 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:15.328 18:59:32 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:15.328 18:59:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:15.328 ************************************ 00:09:15.328 START TEST raid_state_function_test_sb 00:09:15.328 ************************************ 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 3 true 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=78515 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 78515' 00:09:15.328 Process raid pid: 78515 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 78515 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 78515 ']' 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:15.328 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:15.328 18:59:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:15.328 [2024-12-05 18:59:32.736273] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:09:15.328 [2024-12-05 18:59:32.736849] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:15.601 [2024-12-05 18:59:32.890886] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:15.601 [2024-12-05 18:59:32.915583] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:15.601 [2024-12-05 18:59:32.958576] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:15.601 [2024-12-05 18:59:32.958741] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:16.171 18:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:16.171 18:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:09:16.171 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:16.171 18:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.171 18:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:16.171 [2024-12-05 18:59:33.557885] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:16.171 [2024-12-05 18:59:33.558021] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:16.171 [2024-12-05 18:59:33.558053] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:16.171 [2024-12-05 18:59:33.558079] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:16.171 [2024-12-05 18:59:33.558096] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:16.171 [2024-12-05 18:59:33.558119] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:16.171 18:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:16.172 "name": "Existed_Raid", 00:09:16.172 "uuid": "f0b843a2-5f0a-409e-84cb-8064a366d038", 00:09:16.172 "strip_size_kb": 0, 00:09:16.172 "state": "configuring", 00:09:16.172 "raid_level": "raid1", 00:09:16.172 "superblock": true, 00:09:16.172 "num_base_bdevs": 3, 00:09:16.172 "num_base_bdevs_discovered": 0, 00:09:16.172 "num_base_bdevs_operational": 3, 00:09:16.172 "base_bdevs_list": [ 00:09:16.172 { 00:09:16.172 "name": "BaseBdev1", 00:09:16.172 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.172 "is_configured": false, 00:09:16.172 "data_offset": 0, 00:09:16.172 "data_size": 0 00:09:16.172 }, 00:09:16.172 { 00:09:16.172 "name": "BaseBdev2", 00:09:16.172 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.172 "is_configured": false, 00:09:16.172 "data_offset": 0, 00:09:16.172 "data_size": 0 00:09:16.172 }, 00:09:16.172 { 00:09:16.172 "name": "BaseBdev3", 00:09:16.172 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.172 "is_configured": false, 00:09:16.172 "data_offset": 0, 00:09:16.172 "data_size": 0 00:09:16.172 } 00:09:16.172 ] 00:09:16.172 }' 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:16.172 18:59:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:16.743 18:59:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:16.743 [2024-12-05 18:59:34.004979] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:16.743 [2024-12-05 18:59:34.005022] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:16.743 [2024-12-05 18:59:34.016990] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:16.743 [2024-12-05 18:59:34.017033] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:16.743 [2024-12-05 18:59:34.017041] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:16.743 [2024-12-05 18:59:34.017066] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:16.743 [2024-12-05 18:59:34.017072] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:16.743 [2024-12-05 18:59:34.017080] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:16.743 [2024-12-05 18:59:34.037860] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:16.743 BaseBdev1 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:16.743 [ 00:09:16.743 { 00:09:16.743 "name": "BaseBdev1", 00:09:16.743 "aliases": [ 00:09:16.743 "7549ca0b-bcc3-4adb-8fcd-0b616c7270d3" 00:09:16.743 ], 00:09:16.743 "product_name": "Malloc disk", 00:09:16.743 "block_size": 512, 00:09:16.743 "num_blocks": 65536, 00:09:16.743 "uuid": "7549ca0b-bcc3-4adb-8fcd-0b616c7270d3", 00:09:16.743 "assigned_rate_limits": { 00:09:16.743 "rw_ios_per_sec": 0, 00:09:16.743 "rw_mbytes_per_sec": 0, 00:09:16.743 "r_mbytes_per_sec": 0, 00:09:16.743 "w_mbytes_per_sec": 0 00:09:16.743 }, 00:09:16.743 "claimed": true, 00:09:16.743 "claim_type": "exclusive_write", 00:09:16.743 "zoned": false, 00:09:16.743 "supported_io_types": { 00:09:16.743 "read": true, 00:09:16.743 "write": true, 00:09:16.743 "unmap": true, 00:09:16.743 "flush": true, 00:09:16.743 "reset": true, 00:09:16.743 "nvme_admin": false, 00:09:16.743 "nvme_io": false, 00:09:16.743 "nvme_io_md": false, 00:09:16.743 "write_zeroes": true, 00:09:16.743 "zcopy": true, 00:09:16.743 "get_zone_info": false, 00:09:16.743 "zone_management": false, 00:09:16.743 "zone_append": false, 00:09:16.743 "compare": false, 00:09:16.743 "compare_and_write": false, 00:09:16.743 "abort": true, 00:09:16.743 "seek_hole": false, 00:09:16.743 "seek_data": false, 00:09:16.743 "copy": true, 00:09:16.743 "nvme_iov_md": false 00:09:16.743 }, 00:09:16.743 "memory_domains": [ 00:09:16.743 { 00:09:16.743 "dma_device_id": "system", 00:09:16.743 "dma_device_type": 1 00:09:16.743 }, 00:09:16.743 { 00:09:16.743 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:16.743 "dma_device_type": 2 00:09:16.743 } 00:09:16.743 ], 00:09:16.743 "driver_specific": {} 00:09:16.743 } 00:09:16.743 ] 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:16.743 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:16.744 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:16.744 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:16.744 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:16.744 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:16.744 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:16.744 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:16.744 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:16.744 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:16.744 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:16.744 "name": "Existed_Raid", 00:09:16.744 "uuid": "c67e49aa-9086-4024-add9-e21d72f13b24", 00:09:16.744 "strip_size_kb": 0, 00:09:16.744 "state": "configuring", 00:09:16.744 "raid_level": "raid1", 00:09:16.744 "superblock": true, 00:09:16.744 "num_base_bdevs": 3, 00:09:16.744 "num_base_bdevs_discovered": 1, 00:09:16.744 "num_base_bdevs_operational": 3, 00:09:16.744 "base_bdevs_list": [ 00:09:16.744 { 00:09:16.744 "name": "BaseBdev1", 00:09:16.744 "uuid": "7549ca0b-bcc3-4adb-8fcd-0b616c7270d3", 00:09:16.744 "is_configured": true, 00:09:16.744 "data_offset": 2048, 00:09:16.744 "data_size": 63488 00:09:16.744 }, 00:09:16.744 { 00:09:16.744 "name": "BaseBdev2", 00:09:16.744 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.744 "is_configured": false, 00:09:16.744 "data_offset": 0, 00:09:16.744 "data_size": 0 00:09:16.744 }, 00:09:16.744 { 00:09:16.744 "name": "BaseBdev3", 00:09:16.744 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:16.744 "is_configured": false, 00:09:16.744 "data_offset": 0, 00:09:16.744 "data_size": 0 00:09:16.744 } 00:09:16.744 ] 00:09:16.744 }' 00:09:16.744 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:16.744 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:17.005 [2024-12-05 18:59:34.497087] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:17.005 [2024-12-05 18:59:34.497177] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:17.005 [2024-12-05 18:59:34.509111] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:17.005 [2024-12-05 18:59:34.510892] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:17.005 [2024-12-05 18:59:34.510964] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:17.005 [2024-12-05 18:59:34.510976] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:17.005 [2024-12-05 18:59:34.510987] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:17.005 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.265 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:17.265 "name": "Existed_Raid", 00:09:17.265 "uuid": "f032a23d-fd3b-4016-a6d2-78ac92f802c3", 00:09:17.265 "strip_size_kb": 0, 00:09:17.265 "state": "configuring", 00:09:17.265 "raid_level": "raid1", 00:09:17.265 "superblock": true, 00:09:17.265 "num_base_bdevs": 3, 00:09:17.265 "num_base_bdevs_discovered": 1, 00:09:17.265 "num_base_bdevs_operational": 3, 00:09:17.265 "base_bdevs_list": [ 00:09:17.265 { 00:09:17.265 "name": "BaseBdev1", 00:09:17.265 "uuid": "7549ca0b-bcc3-4adb-8fcd-0b616c7270d3", 00:09:17.265 "is_configured": true, 00:09:17.265 "data_offset": 2048, 00:09:17.265 "data_size": 63488 00:09:17.265 }, 00:09:17.265 { 00:09:17.265 "name": "BaseBdev2", 00:09:17.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.265 "is_configured": false, 00:09:17.265 "data_offset": 0, 00:09:17.265 "data_size": 0 00:09:17.265 }, 00:09:17.265 { 00:09:17.265 "name": "BaseBdev3", 00:09:17.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.265 "is_configured": false, 00:09:17.265 "data_offset": 0, 00:09:17.265 "data_size": 0 00:09:17.265 } 00:09:17.265 ] 00:09:17.265 }' 00:09:17.265 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:17.265 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:17.525 [2024-12-05 18:59:34.979499] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:17.525 BaseBdev2 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.525 18:59:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:17.525 [ 00:09:17.525 { 00:09:17.525 "name": "BaseBdev2", 00:09:17.525 "aliases": [ 00:09:17.525 "8d943b51-c046-45b2-859c-79578cd2246f" 00:09:17.525 ], 00:09:17.525 "product_name": "Malloc disk", 00:09:17.525 "block_size": 512, 00:09:17.525 "num_blocks": 65536, 00:09:17.525 "uuid": "8d943b51-c046-45b2-859c-79578cd2246f", 00:09:17.525 "assigned_rate_limits": { 00:09:17.525 "rw_ios_per_sec": 0, 00:09:17.525 "rw_mbytes_per_sec": 0, 00:09:17.525 "r_mbytes_per_sec": 0, 00:09:17.525 "w_mbytes_per_sec": 0 00:09:17.525 }, 00:09:17.525 "claimed": true, 00:09:17.525 "claim_type": "exclusive_write", 00:09:17.525 "zoned": false, 00:09:17.525 "supported_io_types": { 00:09:17.525 "read": true, 00:09:17.525 "write": true, 00:09:17.525 "unmap": true, 00:09:17.525 "flush": true, 00:09:17.525 "reset": true, 00:09:17.525 "nvme_admin": false, 00:09:17.525 "nvme_io": false, 00:09:17.525 "nvme_io_md": false, 00:09:17.525 "write_zeroes": true, 00:09:17.525 "zcopy": true, 00:09:17.525 "get_zone_info": false, 00:09:17.525 "zone_management": false, 00:09:17.525 "zone_append": false, 00:09:17.525 "compare": false, 00:09:17.525 "compare_and_write": false, 00:09:17.525 "abort": true, 00:09:17.525 "seek_hole": false, 00:09:17.525 "seek_data": false, 00:09:17.525 "copy": true, 00:09:17.525 "nvme_iov_md": false 00:09:17.525 }, 00:09:17.525 "memory_domains": [ 00:09:17.525 { 00:09:17.525 "dma_device_id": "system", 00:09:17.525 "dma_device_type": 1 00:09:17.525 }, 00:09:17.525 { 00:09:17.525 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:17.525 "dma_device_type": 2 00:09:17.525 } 00:09:17.525 ], 00:09:17.525 "driver_specific": {} 00:09:17.525 } 00:09:17.525 ] 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:17.525 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:17.525 "name": "Existed_Raid", 00:09:17.525 "uuid": "f032a23d-fd3b-4016-a6d2-78ac92f802c3", 00:09:17.525 "strip_size_kb": 0, 00:09:17.525 "state": "configuring", 00:09:17.526 "raid_level": "raid1", 00:09:17.526 "superblock": true, 00:09:17.526 "num_base_bdevs": 3, 00:09:17.526 "num_base_bdevs_discovered": 2, 00:09:17.526 "num_base_bdevs_operational": 3, 00:09:17.526 "base_bdevs_list": [ 00:09:17.526 { 00:09:17.526 "name": "BaseBdev1", 00:09:17.526 "uuid": "7549ca0b-bcc3-4adb-8fcd-0b616c7270d3", 00:09:17.526 "is_configured": true, 00:09:17.526 "data_offset": 2048, 00:09:17.526 "data_size": 63488 00:09:17.526 }, 00:09:17.526 { 00:09:17.526 "name": "BaseBdev2", 00:09:17.526 "uuid": "8d943b51-c046-45b2-859c-79578cd2246f", 00:09:17.526 "is_configured": true, 00:09:17.526 "data_offset": 2048, 00:09:17.526 "data_size": 63488 00:09:17.526 }, 00:09:17.526 { 00:09:17.526 "name": "BaseBdev3", 00:09:17.526 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:17.526 "is_configured": false, 00:09:17.526 "data_offset": 0, 00:09:17.526 "data_size": 0 00:09:17.526 } 00:09:17.526 ] 00:09:17.526 }' 00:09:17.526 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:17.526 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.094 [2024-12-05 18:59:35.470596] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:18.094 [2024-12-05 18:59:35.471138] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:18.094 [2024-12-05 18:59:35.471213] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:18.094 BaseBdev3 00:09:18.094 [2024-12-05 18:59:35.471863] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:18.094 [2024-12-05 18:59:35.472169] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:18.094 [2024-12-05 18:59:35.472207] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.094 [2024-12-05 18:59:35.472486] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.094 [ 00:09:18.094 { 00:09:18.094 "name": "BaseBdev3", 00:09:18.094 "aliases": [ 00:09:18.094 "cc31ecd9-b195-4f67-a64f-f152c7105e39" 00:09:18.094 ], 00:09:18.094 "product_name": "Malloc disk", 00:09:18.094 "block_size": 512, 00:09:18.094 "num_blocks": 65536, 00:09:18.094 "uuid": "cc31ecd9-b195-4f67-a64f-f152c7105e39", 00:09:18.094 "assigned_rate_limits": { 00:09:18.094 "rw_ios_per_sec": 0, 00:09:18.094 "rw_mbytes_per_sec": 0, 00:09:18.094 "r_mbytes_per_sec": 0, 00:09:18.094 "w_mbytes_per_sec": 0 00:09:18.094 }, 00:09:18.094 "claimed": true, 00:09:18.094 "claim_type": "exclusive_write", 00:09:18.094 "zoned": false, 00:09:18.094 "supported_io_types": { 00:09:18.094 "read": true, 00:09:18.094 "write": true, 00:09:18.094 "unmap": true, 00:09:18.094 "flush": true, 00:09:18.094 "reset": true, 00:09:18.094 "nvme_admin": false, 00:09:18.094 "nvme_io": false, 00:09:18.094 "nvme_io_md": false, 00:09:18.094 "write_zeroes": true, 00:09:18.094 "zcopy": true, 00:09:18.094 "get_zone_info": false, 00:09:18.094 "zone_management": false, 00:09:18.094 "zone_append": false, 00:09:18.094 "compare": false, 00:09:18.094 "compare_and_write": false, 00:09:18.094 "abort": true, 00:09:18.094 "seek_hole": false, 00:09:18.094 "seek_data": false, 00:09:18.094 "copy": true, 00:09:18.094 "nvme_iov_md": false 00:09:18.094 }, 00:09:18.094 "memory_domains": [ 00:09:18.094 { 00:09:18.094 "dma_device_id": "system", 00:09:18.094 "dma_device_type": 1 00:09:18.094 }, 00:09:18.094 { 00:09:18.094 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.094 "dma_device_type": 2 00:09:18.094 } 00:09:18.094 ], 00:09:18.094 "driver_specific": {} 00:09:18.094 } 00:09:18.094 ] 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.094 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.094 "name": "Existed_Raid", 00:09:18.094 "uuid": "f032a23d-fd3b-4016-a6d2-78ac92f802c3", 00:09:18.094 "strip_size_kb": 0, 00:09:18.094 "state": "online", 00:09:18.094 "raid_level": "raid1", 00:09:18.094 "superblock": true, 00:09:18.094 "num_base_bdevs": 3, 00:09:18.094 "num_base_bdevs_discovered": 3, 00:09:18.094 "num_base_bdevs_operational": 3, 00:09:18.094 "base_bdevs_list": [ 00:09:18.094 { 00:09:18.094 "name": "BaseBdev1", 00:09:18.094 "uuid": "7549ca0b-bcc3-4adb-8fcd-0b616c7270d3", 00:09:18.095 "is_configured": true, 00:09:18.095 "data_offset": 2048, 00:09:18.095 "data_size": 63488 00:09:18.095 }, 00:09:18.095 { 00:09:18.095 "name": "BaseBdev2", 00:09:18.095 "uuid": "8d943b51-c046-45b2-859c-79578cd2246f", 00:09:18.095 "is_configured": true, 00:09:18.095 "data_offset": 2048, 00:09:18.095 "data_size": 63488 00:09:18.095 }, 00:09:18.095 { 00:09:18.095 "name": "BaseBdev3", 00:09:18.095 "uuid": "cc31ecd9-b195-4f67-a64f-f152c7105e39", 00:09:18.095 "is_configured": true, 00:09:18.095 "data_offset": 2048, 00:09:18.095 "data_size": 63488 00:09:18.095 } 00:09:18.095 ] 00:09:18.095 }' 00:09:18.095 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.095 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.663 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:18.663 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:18.663 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:18.663 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:18.663 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:18.663 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:18.663 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:18.663 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:18.663 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.663 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.663 [2024-12-05 18:59:35.974023] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:18.663 18:59:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.663 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:18.663 "name": "Existed_Raid", 00:09:18.663 "aliases": [ 00:09:18.663 "f032a23d-fd3b-4016-a6d2-78ac92f802c3" 00:09:18.663 ], 00:09:18.663 "product_name": "Raid Volume", 00:09:18.663 "block_size": 512, 00:09:18.663 "num_blocks": 63488, 00:09:18.663 "uuid": "f032a23d-fd3b-4016-a6d2-78ac92f802c3", 00:09:18.663 "assigned_rate_limits": { 00:09:18.663 "rw_ios_per_sec": 0, 00:09:18.664 "rw_mbytes_per_sec": 0, 00:09:18.664 "r_mbytes_per_sec": 0, 00:09:18.664 "w_mbytes_per_sec": 0 00:09:18.664 }, 00:09:18.664 "claimed": false, 00:09:18.664 "zoned": false, 00:09:18.664 "supported_io_types": { 00:09:18.664 "read": true, 00:09:18.664 "write": true, 00:09:18.664 "unmap": false, 00:09:18.664 "flush": false, 00:09:18.664 "reset": true, 00:09:18.664 "nvme_admin": false, 00:09:18.664 "nvme_io": false, 00:09:18.664 "nvme_io_md": false, 00:09:18.664 "write_zeroes": true, 00:09:18.664 "zcopy": false, 00:09:18.664 "get_zone_info": false, 00:09:18.664 "zone_management": false, 00:09:18.664 "zone_append": false, 00:09:18.664 "compare": false, 00:09:18.664 "compare_and_write": false, 00:09:18.664 "abort": false, 00:09:18.664 "seek_hole": false, 00:09:18.664 "seek_data": false, 00:09:18.664 "copy": false, 00:09:18.664 "nvme_iov_md": false 00:09:18.664 }, 00:09:18.664 "memory_domains": [ 00:09:18.664 { 00:09:18.664 "dma_device_id": "system", 00:09:18.664 "dma_device_type": 1 00:09:18.664 }, 00:09:18.664 { 00:09:18.664 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.664 "dma_device_type": 2 00:09:18.664 }, 00:09:18.664 { 00:09:18.664 "dma_device_id": "system", 00:09:18.664 "dma_device_type": 1 00:09:18.664 }, 00:09:18.664 { 00:09:18.664 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.664 "dma_device_type": 2 00:09:18.664 }, 00:09:18.664 { 00:09:18.664 "dma_device_id": "system", 00:09:18.664 "dma_device_type": 1 00:09:18.664 }, 00:09:18.664 { 00:09:18.664 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:18.664 "dma_device_type": 2 00:09:18.664 } 00:09:18.664 ], 00:09:18.664 "driver_specific": { 00:09:18.664 "raid": { 00:09:18.664 "uuid": "f032a23d-fd3b-4016-a6d2-78ac92f802c3", 00:09:18.664 "strip_size_kb": 0, 00:09:18.664 "state": "online", 00:09:18.664 "raid_level": "raid1", 00:09:18.664 "superblock": true, 00:09:18.664 "num_base_bdevs": 3, 00:09:18.664 "num_base_bdevs_discovered": 3, 00:09:18.664 "num_base_bdevs_operational": 3, 00:09:18.664 "base_bdevs_list": [ 00:09:18.664 { 00:09:18.664 "name": "BaseBdev1", 00:09:18.664 "uuid": "7549ca0b-bcc3-4adb-8fcd-0b616c7270d3", 00:09:18.664 "is_configured": true, 00:09:18.664 "data_offset": 2048, 00:09:18.664 "data_size": 63488 00:09:18.664 }, 00:09:18.664 { 00:09:18.664 "name": "BaseBdev2", 00:09:18.664 "uuid": "8d943b51-c046-45b2-859c-79578cd2246f", 00:09:18.664 "is_configured": true, 00:09:18.664 "data_offset": 2048, 00:09:18.664 "data_size": 63488 00:09:18.664 }, 00:09:18.664 { 00:09:18.664 "name": "BaseBdev3", 00:09:18.664 "uuid": "cc31ecd9-b195-4f67-a64f-f152c7105e39", 00:09:18.664 "is_configured": true, 00:09:18.664 "data_offset": 2048, 00:09:18.664 "data_size": 63488 00:09:18.664 } 00:09:18.664 ] 00:09:18.664 } 00:09:18.664 } 00:09:18.664 }' 00:09:18.664 18:59:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:18.664 BaseBdev2 00:09:18.664 BaseBdev3' 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.664 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.664 [2024-12-05 18:59:36.209371] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:18.922 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.922 "name": "Existed_Raid", 00:09:18.922 "uuid": "f032a23d-fd3b-4016-a6d2-78ac92f802c3", 00:09:18.922 "strip_size_kb": 0, 00:09:18.922 "state": "online", 00:09:18.922 "raid_level": "raid1", 00:09:18.922 "superblock": true, 00:09:18.922 "num_base_bdevs": 3, 00:09:18.922 "num_base_bdevs_discovered": 2, 00:09:18.922 "num_base_bdevs_operational": 2, 00:09:18.922 "base_bdevs_list": [ 00:09:18.922 { 00:09:18.922 "name": null, 00:09:18.922 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:18.922 "is_configured": false, 00:09:18.922 "data_offset": 0, 00:09:18.922 "data_size": 63488 00:09:18.922 }, 00:09:18.922 { 00:09:18.922 "name": "BaseBdev2", 00:09:18.922 "uuid": "8d943b51-c046-45b2-859c-79578cd2246f", 00:09:18.922 "is_configured": true, 00:09:18.922 "data_offset": 2048, 00:09:18.922 "data_size": 63488 00:09:18.922 }, 00:09:18.922 { 00:09:18.922 "name": "BaseBdev3", 00:09:18.922 "uuid": "cc31ecd9-b195-4f67-a64f-f152c7105e39", 00:09:18.923 "is_configured": true, 00:09:18.923 "data_offset": 2048, 00:09:18.923 "data_size": 63488 00:09:18.923 } 00:09:18.923 ] 00:09:18.923 }' 00:09:18.923 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.923 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.181 [2024-12-05 18:59:36.663814] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.181 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.181 [2024-12-05 18:59:36.734833] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:19.181 [2024-12-05 18:59:36.734926] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:19.440 [2024-12-05 18:59:36.746409] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:19.440 [2024-12-05 18:59:36.746558] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:19.440 [2024-12-05 18:59:36.746578] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.440 BaseBdev2 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.440 [ 00:09:19.440 { 00:09:19.440 "name": "BaseBdev2", 00:09:19.440 "aliases": [ 00:09:19.440 "ff5f7694-d42a-40aa-bb13-44d6416ad236" 00:09:19.440 ], 00:09:19.440 "product_name": "Malloc disk", 00:09:19.440 "block_size": 512, 00:09:19.440 "num_blocks": 65536, 00:09:19.440 "uuid": "ff5f7694-d42a-40aa-bb13-44d6416ad236", 00:09:19.440 "assigned_rate_limits": { 00:09:19.440 "rw_ios_per_sec": 0, 00:09:19.440 "rw_mbytes_per_sec": 0, 00:09:19.440 "r_mbytes_per_sec": 0, 00:09:19.440 "w_mbytes_per_sec": 0 00:09:19.440 }, 00:09:19.440 "claimed": false, 00:09:19.440 "zoned": false, 00:09:19.440 "supported_io_types": { 00:09:19.440 "read": true, 00:09:19.440 "write": true, 00:09:19.440 "unmap": true, 00:09:19.440 "flush": true, 00:09:19.440 "reset": true, 00:09:19.440 "nvme_admin": false, 00:09:19.440 "nvme_io": false, 00:09:19.440 "nvme_io_md": false, 00:09:19.440 "write_zeroes": true, 00:09:19.440 "zcopy": true, 00:09:19.440 "get_zone_info": false, 00:09:19.440 "zone_management": false, 00:09:19.440 "zone_append": false, 00:09:19.440 "compare": false, 00:09:19.440 "compare_and_write": false, 00:09:19.440 "abort": true, 00:09:19.440 "seek_hole": false, 00:09:19.440 "seek_data": false, 00:09:19.440 "copy": true, 00:09:19.440 "nvme_iov_md": false 00:09:19.440 }, 00:09:19.440 "memory_domains": [ 00:09:19.440 { 00:09:19.440 "dma_device_id": "system", 00:09:19.440 "dma_device_type": 1 00:09:19.440 }, 00:09:19.440 { 00:09:19.440 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:19.440 "dma_device_type": 2 00:09:19.440 } 00:09:19.440 ], 00:09:19.440 "driver_specific": {} 00:09:19.440 } 00:09:19.440 ] 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.440 BaseBdev3 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.440 [ 00:09:19.440 { 00:09:19.440 "name": "BaseBdev3", 00:09:19.440 "aliases": [ 00:09:19.440 "fafe98eb-8057-4fd1-a9f4-5dd2bfdafe77" 00:09:19.440 ], 00:09:19.440 "product_name": "Malloc disk", 00:09:19.440 "block_size": 512, 00:09:19.440 "num_blocks": 65536, 00:09:19.440 "uuid": "fafe98eb-8057-4fd1-a9f4-5dd2bfdafe77", 00:09:19.440 "assigned_rate_limits": { 00:09:19.440 "rw_ios_per_sec": 0, 00:09:19.440 "rw_mbytes_per_sec": 0, 00:09:19.440 "r_mbytes_per_sec": 0, 00:09:19.440 "w_mbytes_per_sec": 0 00:09:19.440 }, 00:09:19.440 "claimed": false, 00:09:19.440 "zoned": false, 00:09:19.440 "supported_io_types": { 00:09:19.440 "read": true, 00:09:19.440 "write": true, 00:09:19.440 "unmap": true, 00:09:19.440 "flush": true, 00:09:19.440 "reset": true, 00:09:19.440 "nvme_admin": false, 00:09:19.440 "nvme_io": false, 00:09:19.440 "nvme_io_md": false, 00:09:19.440 "write_zeroes": true, 00:09:19.440 "zcopy": true, 00:09:19.440 "get_zone_info": false, 00:09:19.440 "zone_management": false, 00:09:19.440 "zone_append": false, 00:09:19.440 "compare": false, 00:09:19.440 "compare_and_write": false, 00:09:19.440 "abort": true, 00:09:19.440 "seek_hole": false, 00:09:19.440 "seek_data": false, 00:09:19.440 "copy": true, 00:09:19.440 "nvme_iov_md": false 00:09:19.440 }, 00:09:19.440 "memory_domains": [ 00:09:19.440 { 00:09:19.440 "dma_device_id": "system", 00:09:19.440 "dma_device_type": 1 00:09:19.440 }, 00:09:19.440 { 00:09:19.440 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:19.440 "dma_device_type": 2 00:09:19.440 } 00:09:19.440 ], 00:09:19.440 "driver_specific": {} 00:09:19.440 } 00:09:19.440 ] 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.440 [2024-12-05 18:59:36.909773] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:19.440 [2024-12-05 18:59:36.909871] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:19.440 [2024-12-05 18:59:36.909916] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:19.440 [2024-12-05 18:59:36.911796] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:19.440 "name": "Existed_Raid", 00:09:19.440 "uuid": "23ba8447-9705-4757-a9b7-6f9fae4656d4", 00:09:19.440 "strip_size_kb": 0, 00:09:19.440 "state": "configuring", 00:09:19.440 "raid_level": "raid1", 00:09:19.440 "superblock": true, 00:09:19.440 "num_base_bdevs": 3, 00:09:19.440 "num_base_bdevs_discovered": 2, 00:09:19.440 "num_base_bdevs_operational": 3, 00:09:19.440 "base_bdevs_list": [ 00:09:19.440 { 00:09:19.440 "name": "BaseBdev1", 00:09:19.440 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.440 "is_configured": false, 00:09:19.440 "data_offset": 0, 00:09:19.440 "data_size": 0 00:09:19.440 }, 00:09:19.440 { 00:09:19.440 "name": "BaseBdev2", 00:09:19.440 "uuid": "ff5f7694-d42a-40aa-bb13-44d6416ad236", 00:09:19.440 "is_configured": true, 00:09:19.440 "data_offset": 2048, 00:09:19.440 "data_size": 63488 00:09:19.440 }, 00:09:19.440 { 00:09:19.440 "name": "BaseBdev3", 00:09:19.440 "uuid": "fafe98eb-8057-4fd1-a9f4-5dd2bfdafe77", 00:09:19.440 "is_configured": true, 00:09:19.440 "data_offset": 2048, 00:09:19.440 "data_size": 63488 00:09:19.440 } 00:09:19.440 ] 00:09:19.440 }' 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:19.440 18:59:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.009 [2024-12-05 18:59:37.349010] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:20.009 "name": "Existed_Raid", 00:09:20.009 "uuid": "23ba8447-9705-4757-a9b7-6f9fae4656d4", 00:09:20.009 "strip_size_kb": 0, 00:09:20.009 "state": "configuring", 00:09:20.009 "raid_level": "raid1", 00:09:20.009 "superblock": true, 00:09:20.009 "num_base_bdevs": 3, 00:09:20.009 "num_base_bdevs_discovered": 1, 00:09:20.009 "num_base_bdevs_operational": 3, 00:09:20.009 "base_bdevs_list": [ 00:09:20.009 { 00:09:20.009 "name": "BaseBdev1", 00:09:20.009 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:20.009 "is_configured": false, 00:09:20.009 "data_offset": 0, 00:09:20.009 "data_size": 0 00:09:20.009 }, 00:09:20.009 { 00:09:20.009 "name": null, 00:09:20.009 "uuid": "ff5f7694-d42a-40aa-bb13-44d6416ad236", 00:09:20.009 "is_configured": false, 00:09:20.009 "data_offset": 0, 00:09:20.009 "data_size": 63488 00:09:20.009 }, 00:09:20.009 { 00:09:20.009 "name": "BaseBdev3", 00:09:20.009 "uuid": "fafe98eb-8057-4fd1-a9f4-5dd2bfdafe77", 00:09:20.009 "is_configured": true, 00:09:20.009 "data_offset": 2048, 00:09:20.009 "data_size": 63488 00:09:20.009 } 00:09:20.009 ] 00:09:20.009 }' 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:20.009 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.269 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.269 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.270 [2024-12-05 18:59:37.803389] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:20.270 BaseBdev1 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.270 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.530 [ 00:09:20.530 { 00:09:20.530 "name": "BaseBdev1", 00:09:20.530 "aliases": [ 00:09:20.530 "6837e5b6-be13-40d3-a80d-b30ca14ba5b6" 00:09:20.530 ], 00:09:20.530 "product_name": "Malloc disk", 00:09:20.530 "block_size": 512, 00:09:20.530 "num_blocks": 65536, 00:09:20.530 "uuid": "6837e5b6-be13-40d3-a80d-b30ca14ba5b6", 00:09:20.530 "assigned_rate_limits": { 00:09:20.530 "rw_ios_per_sec": 0, 00:09:20.530 "rw_mbytes_per_sec": 0, 00:09:20.530 "r_mbytes_per_sec": 0, 00:09:20.530 "w_mbytes_per_sec": 0 00:09:20.530 }, 00:09:20.530 "claimed": true, 00:09:20.530 "claim_type": "exclusive_write", 00:09:20.530 "zoned": false, 00:09:20.530 "supported_io_types": { 00:09:20.530 "read": true, 00:09:20.530 "write": true, 00:09:20.530 "unmap": true, 00:09:20.530 "flush": true, 00:09:20.530 "reset": true, 00:09:20.530 "nvme_admin": false, 00:09:20.530 "nvme_io": false, 00:09:20.530 "nvme_io_md": false, 00:09:20.530 "write_zeroes": true, 00:09:20.530 "zcopy": true, 00:09:20.530 "get_zone_info": false, 00:09:20.530 "zone_management": false, 00:09:20.530 "zone_append": false, 00:09:20.530 "compare": false, 00:09:20.530 "compare_and_write": false, 00:09:20.530 "abort": true, 00:09:20.530 "seek_hole": false, 00:09:20.530 "seek_data": false, 00:09:20.530 "copy": true, 00:09:20.530 "nvme_iov_md": false 00:09:20.530 }, 00:09:20.530 "memory_domains": [ 00:09:20.530 { 00:09:20.530 "dma_device_id": "system", 00:09:20.530 "dma_device_type": 1 00:09:20.530 }, 00:09:20.530 { 00:09:20.530 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:20.530 "dma_device_type": 2 00:09:20.530 } 00:09:20.530 ], 00:09:20.530 "driver_specific": {} 00:09:20.530 } 00:09:20.530 ] 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.530 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:20.530 "name": "Existed_Raid", 00:09:20.530 "uuid": "23ba8447-9705-4757-a9b7-6f9fae4656d4", 00:09:20.530 "strip_size_kb": 0, 00:09:20.530 "state": "configuring", 00:09:20.530 "raid_level": "raid1", 00:09:20.530 "superblock": true, 00:09:20.530 "num_base_bdevs": 3, 00:09:20.530 "num_base_bdevs_discovered": 2, 00:09:20.530 "num_base_bdevs_operational": 3, 00:09:20.530 "base_bdevs_list": [ 00:09:20.530 { 00:09:20.530 "name": "BaseBdev1", 00:09:20.530 "uuid": "6837e5b6-be13-40d3-a80d-b30ca14ba5b6", 00:09:20.530 "is_configured": true, 00:09:20.530 "data_offset": 2048, 00:09:20.530 "data_size": 63488 00:09:20.530 }, 00:09:20.530 { 00:09:20.530 "name": null, 00:09:20.531 "uuid": "ff5f7694-d42a-40aa-bb13-44d6416ad236", 00:09:20.531 "is_configured": false, 00:09:20.531 "data_offset": 0, 00:09:20.531 "data_size": 63488 00:09:20.531 }, 00:09:20.531 { 00:09:20.531 "name": "BaseBdev3", 00:09:20.531 "uuid": "fafe98eb-8057-4fd1-a9f4-5dd2bfdafe77", 00:09:20.531 "is_configured": true, 00:09:20.531 "data_offset": 2048, 00:09:20.531 "data_size": 63488 00:09:20.531 } 00:09:20.531 ] 00:09:20.531 }' 00:09:20.531 18:59:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:20.531 18:59:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.791 [2024-12-05 18:59:38.262645] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:20.791 "name": "Existed_Raid", 00:09:20.791 "uuid": "23ba8447-9705-4757-a9b7-6f9fae4656d4", 00:09:20.791 "strip_size_kb": 0, 00:09:20.791 "state": "configuring", 00:09:20.791 "raid_level": "raid1", 00:09:20.791 "superblock": true, 00:09:20.791 "num_base_bdevs": 3, 00:09:20.791 "num_base_bdevs_discovered": 1, 00:09:20.791 "num_base_bdevs_operational": 3, 00:09:20.791 "base_bdevs_list": [ 00:09:20.791 { 00:09:20.791 "name": "BaseBdev1", 00:09:20.791 "uuid": "6837e5b6-be13-40d3-a80d-b30ca14ba5b6", 00:09:20.791 "is_configured": true, 00:09:20.791 "data_offset": 2048, 00:09:20.791 "data_size": 63488 00:09:20.791 }, 00:09:20.791 { 00:09:20.791 "name": null, 00:09:20.791 "uuid": "ff5f7694-d42a-40aa-bb13-44d6416ad236", 00:09:20.791 "is_configured": false, 00:09:20.791 "data_offset": 0, 00:09:20.791 "data_size": 63488 00:09:20.791 }, 00:09:20.791 { 00:09:20.791 "name": null, 00:09:20.791 "uuid": "fafe98eb-8057-4fd1-a9f4-5dd2bfdafe77", 00:09:20.791 "is_configured": false, 00:09:20.791 "data_offset": 0, 00:09:20.791 "data_size": 63488 00:09:20.791 } 00:09:20.791 ] 00:09:20.791 }' 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:20.791 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.361 [2024-12-05 18:59:38.733835] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:21.361 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:21.362 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:21.362 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:21.362 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:21.362 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.362 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.362 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.362 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:21.362 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.362 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:21.362 "name": "Existed_Raid", 00:09:21.362 "uuid": "23ba8447-9705-4757-a9b7-6f9fae4656d4", 00:09:21.362 "strip_size_kb": 0, 00:09:21.362 "state": "configuring", 00:09:21.362 "raid_level": "raid1", 00:09:21.362 "superblock": true, 00:09:21.362 "num_base_bdevs": 3, 00:09:21.362 "num_base_bdevs_discovered": 2, 00:09:21.362 "num_base_bdevs_operational": 3, 00:09:21.362 "base_bdevs_list": [ 00:09:21.362 { 00:09:21.362 "name": "BaseBdev1", 00:09:21.362 "uuid": "6837e5b6-be13-40d3-a80d-b30ca14ba5b6", 00:09:21.362 "is_configured": true, 00:09:21.362 "data_offset": 2048, 00:09:21.362 "data_size": 63488 00:09:21.362 }, 00:09:21.362 { 00:09:21.362 "name": null, 00:09:21.362 "uuid": "ff5f7694-d42a-40aa-bb13-44d6416ad236", 00:09:21.362 "is_configured": false, 00:09:21.362 "data_offset": 0, 00:09:21.362 "data_size": 63488 00:09:21.362 }, 00:09:21.362 { 00:09:21.362 "name": "BaseBdev3", 00:09:21.362 "uuid": "fafe98eb-8057-4fd1-a9f4-5dd2bfdafe77", 00:09:21.362 "is_configured": true, 00:09:21.362 "data_offset": 2048, 00:09:21.362 "data_size": 63488 00:09:21.362 } 00:09:21.362 ] 00:09:21.362 }' 00:09:21.362 18:59:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:21.362 18:59:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.622 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.622 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.622 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.622 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:21.622 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.883 [2024-12-05 18:59:39.221028] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:21.883 "name": "Existed_Raid", 00:09:21.883 "uuid": "23ba8447-9705-4757-a9b7-6f9fae4656d4", 00:09:21.883 "strip_size_kb": 0, 00:09:21.883 "state": "configuring", 00:09:21.883 "raid_level": "raid1", 00:09:21.883 "superblock": true, 00:09:21.883 "num_base_bdevs": 3, 00:09:21.883 "num_base_bdevs_discovered": 1, 00:09:21.883 "num_base_bdevs_operational": 3, 00:09:21.883 "base_bdevs_list": [ 00:09:21.883 { 00:09:21.883 "name": null, 00:09:21.883 "uuid": "6837e5b6-be13-40d3-a80d-b30ca14ba5b6", 00:09:21.883 "is_configured": false, 00:09:21.883 "data_offset": 0, 00:09:21.883 "data_size": 63488 00:09:21.883 }, 00:09:21.883 { 00:09:21.883 "name": null, 00:09:21.883 "uuid": "ff5f7694-d42a-40aa-bb13-44d6416ad236", 00:09:21.883 "is_configured": false, 00:09:21.883 "data_offset": 0, 00:09:21.883 "data_size": 63488 00:09:21.883 }, 00:09:21.883 { 00:09:21.883 "name": "BaseBdev3", 00:09:21.883 "uuid": "fafe98eb-8057-4fd1-a9f4-5dd2bfdafe77", 00:09:21.883 "is_configured": true, 00:09:21.883 "data_offset": 2048, 00:09:21.883 "data_size": 63488 00:09:21.883 } 00:09:21.883 ] 00:09:21.883 }' 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:21.883 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.143 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.143 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:22.143 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.143 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.143 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.403 [2024-12-05 18:59:39.718559] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.403 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.403 "name": "Existed_Raid", 00:09:22.403 "uuid": "23ba8447-9705-4757-a9b7-6f9fae4656d4", 00:09:22.403 "strip_size_kb": 0, 00:09:22.403 "state": "configuring", 00:09:22.403 "raid_level": "raid1", 00:09:22.403 "superblock": true, 00:09:22.403 "num_base_bdevs": 3, 00:09:22.404 "num_base_bdevs_discovered": 2, 00:09:22.404 "num_base_bdevs_operational": 3, 00:09:22.404 "base_bdevs_list": [ 00:09:22.404 { 00:09:22.404 "name": null, 00:09:22.404 "uuid": "6837e5b6-be13-40d3-a80d-b30ca14ba5b6", 00:09:22.404 "is_configured": false, 00:09:22.404 "data_offset": 0, 00:09:22.404 "data_size": 63488 00:09:22.404 }, 00:09:22.404 { 00:09:22.404 "name": "BaseBdev2", 00:09:22.404 "uuid": "ff5f7694-d42a-40aa-bb13-44d6416ad236", 00:09:22.404 "is_configured": true, 00:09:22.404 "data_offset": 2048, 00:09:22.404 "data_size": 63488 00:09:22.404 }, 00:09:22.404 { 00:09:22.404 "name": "BaseBdev3", 00:09:22.404 "uuid": "fafe98eb-8057-4fd1-a9f4-5dd2bfdafe77", 00:09:22.404 "is_configured": true, 00:09:22.404 "data_offset": 2048, 00:09:22.404 "data_size": 63488 00:09:22.404 } 00:09:22.404 ] 00:09:22.404 }' 00:09:22.404 18:59:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.404 18:59:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.665 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.665 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.665 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.665 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:22.665 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.665 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:22.665 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.665 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.665 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.665 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:22.665 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 6837e5b6-be13-40d3-a80d-b30ca14ba5b6 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.925 [2024-12-05 18:59:40.240738] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:22.925 [2024-12-05 18:59:40.240908] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:22.925 [2024-12-05 18:59:40.240921] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:22.925 [2024-12-05 18:59:40.241176] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:09:22.925 [2024-12-05 18:59:40.241286] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:22.925 [2024-12-05 18:59:40.241299] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:22.925 [2024-12-05 18:59:40.241391] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:22.925 NewBaseBdev 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.925 [ 00:09:22.925 { 00:09:22.925 "name": "NewBaseBdev", 00:09:22.925 "aliases": [ 00:09:22.925 "6837e5b6-be13-40d3-a80d-b30ca14ba5b6" 00:09:22.925 ], 00:09:22.925 "product_name": "Malloc disk", 00:09:22.925 "block_size": 512, 00:09:22.925 "num_blocks": 65536, 00:09:22.925 "uuid": "6837e5b6-be13-40d3-a80d-b30ca14ba5b6", 00:09:22.925 "assigned_rate_limits": { 00:09:22.925 "rw_ios_per_sec": 0, 00:09:22.925 "rw_mbytes_per_sec": 0, 00:09:22.925 "r_mbytes_per_sec": 0, 00:09:22.925 "w_mbytes_per_sec": 0 00:09:22.925 }, 00:09:22.925 "claimed": true, 00:09:22.925 "claim_type": "exclusive_write", 00:09:22.925 "zoned": false, 00:09:22.925 "supported_io_types": { 00:09:22.925 "read": true, 00:09:22.925 "write": true, 00:09:22.925 "unmap": true, 00:09:22.925 "flush": true, 00:09:22.925 "reset": true, 00:09:22.925 "nvme_admin": false, 00:09:22.925 "nvme_io": false, 00:09:22.925 "nvme_io_md": false, 00:09:22.925 "write_zeroes": true, 00:09:22.925 "zcopy": true, 00:09:22.925 "get_zone_info": false, 00:09:22.925 "zone_management": false, 00:09:22.925 "zone_append": false, 00:09:22.925 "compare": false, 00:09:22.925 "compare_and_write": false, 00:09:22.925 "abort": true, 00:09:22.925 "seek_hole": false, 00:09:22.925 "seek_data": false, 00:09:22.925 "copy": true, 00:09:22.925 "nvme_iov_md": false 00:09:22.925 }, 00:09:22.925 "memory_domains": [ 00:09:22.925 { 00:09:22.925 "dma_device_id": "system", 00:09:22.925 "dma_device_type": 1 00:09:22.925 }, 00:09:22.925 { 00:09:22.925 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:22.925 "dma_device_type": 2 00:09:22.925 } 00:09:22.925 ], 00:09:22.925 "driver_specific": {} 00:09:22.925 } 00:09:22.925 ] 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.925 "name": "Existed_Raid", 00:09:22.925 "uuid": "23ba8447-9705-4757-a9b7-6f9fae4656d4", 00:09:22.925 "strip_size_kb": 0, 00:09:22.925 "state": "online", 00:09:22.925 "raid_level": "raid1", 00:09:22.925 "superblock": true, 00:09:22.925 "num_base_bdevs": 3, 00:09:22.925 "num_base_bdevs_discovered": 3, 00:09:22.925 "num_base_bdevs_operational": 3, 00:09:22.925 "base_bdevs_list": [ 00:09:22.925 { 00:09:22.925 "name": "NewBaseBdev", 00:09:22.925 "uuid": "6837e5b6-be13-40d3-a80d-b30ca14ba5b6", 00:09:22.925 "is_configured": true, 00:09:22.925 "data_offset": 2048, 00:09:22.925 "data_size": 63488 00:09:22.925 }, 00:09:22.925 { 00:09:22.925 "name": "BaseBdev2", 00:09:22.925 "uuid": "ff5f7694-d42a-40aa-bb13-44d6416ad236", 00:09:22.925 "is_configured": true, 00:09:22.925 "data_offset": 2048, 00:09:22.925 "data_size": 63488 00:09:22.925 }, 00:09:22.925 { 00:09:22.925 "name": "BaseBdev3", 00:09:22.925 "uuid": "fafe98eb-8057-4fd1-a9f4-5dd2bfdafe77", 00:09:22.925 "is_configured": true, 00:09:22.925 "data_offset": 2048, 00:09:22.925 "data_size": 63488 00:09:22.925 } 00:09:22.925 ] 00:09:22.925 }' 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.925 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.184 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:23.184 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:23.184 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:23.185 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:23.444 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:23.444 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:23.444 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:23.444 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:23.444 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.444 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.444 [2024-12-05 18:59:40.752278] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:23.444 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.444 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:23.444 "name": "Existed_Raid", 00:09:23.444 "aliases": [ 00:09:23.444 "23ba8447-9705-4757-a9b7-6f9fae4656d4" 00:09:23.444 ], 00:09:23.444 "product_name": "Raid Volume", 00:09:23.444 "block_size": 512, 00:09:23.444 "num_blocks": 63488, 00:09:23.444 "uuid": "23ba8447-9705-4757-a9b7-6f9fae4656d4", 00:09:23.444 "assigned_rate_limits": { 00:09:23.444 "rw_ios_per_sec": 0, 00:09:23.444 "rw_mbytes_per_sec": 0, 00:09:23.444 "r_mbytes_per_sec": 0, 00:09:23.444 "w_mbytes_per_sec": 0 00:09:23.444 }, 00:09:23.444 "claimed": false, 00:09:23.444 "zoned": false, 00:09:23.444 "supported_io_types": { 00:09:23.444 "read": true, 00:09:23.444 "write": true, 00:09:23.444 "unmap": false, 00:09:23.444 "flush": false, 00:09:23.444 "reset": true, 00:09:23.444 "nvme_admin": false, 00:09:23.444 "nvme_io": false, 00:09:23.444 "nvme_io_md": false, 00:09:23.444 "write_zeroes": true, 00:09:23.444 "zcopy": false, 00:09:23.444 "get_zone_info": false, 00:09:23.444 "zone_management": false, 00:09:23.444 "zone_append": false, 00:09:23.444 "compare": false, 00:09:23.444 "compare_and_write": false, 00:09:23.444 "abort": false, 00:09:23.444 "seek_hole": false, 00:09:23.444 "seek_data": false, 00:09:23.444 "copy": false, 00:09:23.444 "nvme_iov_md": false 00:09:23.444 }, 00:09:23.444 "memory_domains": [ 00:09:23.444 { 00:09:23.444 "dma_device_id": "system", 00:09:23.444 "dma_device_type": 1 00:09:23.444 }, 00:09:23.444 { 00:09:23.444 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:23.444 "dma_device_type": 2 00:09:23.444 }, 00:09:23.444 { 00:09:23.444 "dma_device_id": "system", 00:09:23.444 "dma_device_type": 1 00:09:23.444 }, 00:09:23.444 { 00:09:23.444 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:23.444 "dma_device_type": 2 00:09:23.444 }, 00:09:23.444 { 00:09:23.444 "dma_device_id": "system", 00:09:23.444 "dma_device_type": 1 00:09:23.444 }, 00:09:23.444 { 00:09:23.444 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:23.444 "dma_device_type": 2 00:09:23.444 } 00:09:23.444 ], 00:09:23.444 "driver_specific": { 00:09:23.444 "raid": { 00:09:23.444 "uuid": "23ba8447-9705-4757-a9b7-6f9fae4656d4", 00:09:23.444 "strip_size_kb": 0, 00:09:23.444 "state": "online", 00:09:23.444 "raid_level": "raid1", 00:09:23.444 "superblock": true, 00:09:23.444 "num_base_bdevs": 3, 00:09:23.444 "num_base_bdevs_discovered": 3, 00:09:23.444 "num_base_bdevs_operational": 3, 00:09:23.444 "base_bdevs_list": [ 00:09:23.444 { 00:09:23.444 "name": "NewBaseBdev", 00:09:23.444 "uuid": "6837e5b6-be13-40d3-a80d-b30ca14ba5b6", 00:09:23.445 "is_configured": true, 00:09:23.445 "data_offset": 2048, 00:09:23.445 "data_size": 63488 00:09:23.445 }, 00:09:23.445 { 00:09:23.445 "name": "BaseBdev2", 00:09:23.445 "uuid": "ff5f7694-d42a-40aa-bb13-44d6416ad236", 00:09:23.445 "is_configured": true, 00:09:23.445 "data_offset": 2048, 00:09:23.445 "data_size": 63488 00:09:23.445 }, 00:09:23.445 { 00:09:23.445 "name": "BaseBdev3", 00:09:23.445 "uuid": "fafe98eb-8057-4fd1-a9f4-5dd2bfdafe77", 00:09:23.445 "is_configured": true, 00:09:23.445 "data_offset": 2048, 00:09:23.445 "data_size": 63488 00:09:23.445 } 00:09:23.445 ] 00:09:23.445 } 00:09:23.445 } 00:09:23.445 }' 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:23.445 BaseBdev2 00:09:23.445 BaseBdev3' 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.445 18:59:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.704 [2024-12-05 18:59:41.023586] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:23.704 [2024-12-05 18:59:41.023655] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:23.704 [2024-12-05 18:59:41.023748] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:23.704 [2024-12-05 18:59:41.024012] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:23.704 [2024-12-05 18:59:41.024069] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 78515 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 78515 ']' 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 78515 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 78515 00:09:23.704 killing process with pid 78515 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 78515' 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 78515 00:09:23.704 [2024-12-05 18:59:41.070476] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:23.704 18:59:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 78515 00:09:23.704 [2024-12-05 18:59:41.100687] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:23.964 18:59:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:23.964 00:09:23.964 real 0m8.671s 00:09:23.964 user 0m14.880s 00:09:23.964 sys 0m1.678s 00:09:23.964 18:59:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:23.964 ************************************ 00:09:23.964 END TEST raid_state_function_test_sb 00:09:23.964 ************************************ 00:09:23.964 18:59:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:23.964 18:59:41 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 3 00:09:23.964 18:59:41 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:09:23.964 18:59:41 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:23.964 18:59:41 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:23.964 ************************************ 00:09:23.964 START TEST raid_superblock_test 00:09:23.964 ************************************ 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 3 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=79066 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 79066 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 79066 ']' 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:23.964 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:23.964 18:59:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.964 [2024-12-05 18:59:41.472419] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:09:23.964 [2024-12-05 18:59:41.472541] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79066 ] 00:09:24.223 [2024-12-05 18:59:41.620873] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:24.223 [2024-12-05 18:59:41.645492] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:24.223 [2024-12-05 18:59:41.688408] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:24.223 [2024-12-05 18:59:41.688535] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.792 malloc1 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.792 [2024-12-05 18:59:42.316592] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:24.792 [2024-12-05 18:59:42.316712] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:24.792 [2024-12-05 18:59:42.316755] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:09:24.792 [2024-12-05 18:59:42.316801] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:24.792 [2024-12-05 18:59:42.318961] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:24.792 [2024-12-05 18:59:42.319037] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:24.792 pt1 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.792 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:24.793 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:24.793 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:24.793 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:24.793 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:24.793 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:24.793 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:24.793 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:24.793 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:24.793 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.793 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.793 malloc2 00:09:24.793 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:24.793 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:24.793 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:24.793 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.793 [2024-12-05 18:59:42.349444] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:24.793 [2024-12-05 18:59:42.349560] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:24.793 [2024-12-05 18:59:42.349598] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:24.793 [2024-12-05 18:59:42.349630] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:25.052 [2024-12-05 18:59:42.351793] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:25.052 [2024-12-05 18:59:42.351866] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:25.052 pt2 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.052 malloc3 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.052 [2024-12-05 18:59:42.374251] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:25.052 [2024-12-05 18:59:42.374360] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:25.052 [2024-12-05 18:59:42.374395] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:25.052 [2024-12-05 18:59:42.374430] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:25.052 [2024-12-05 18:59:42.376534] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:25.052 [2024-12-05 18:59:42.376612] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:25.052 pt3 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.052 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.052 [2024-12-05 18:59:42.382297] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:25.052 [2024-12-05 18:59:42.384179] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:25.053 [2024-12-05 18:59:42.384277] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:25.053 [2024-12-05 18:59:42.384460] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:09:25.053 [2024-12-05 18:59:42.384506] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:25.053 [2024-12-05 18:59:42.384782] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:25.053 [2024-12-05 18:59:42.384956] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:09:25.053 [2024-12-05 18:59:42.385002] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:09:25.053 [2024-12-05 18:59:42.385144] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:25.053 "name": "raid_bdev1", 00:09:25.053 "uuid": "0ed75771-455d-479f-ad9f-99c5bdb6d298", 00:09:25.053 "strip_size_kb": 0, 00:09:25.053 "state": "online", 00:09:25.053 "raid_level": "raid1", 00:09:25.053 "superblock": true, 00:09:25.053 "num_base_bdevs": 3, 00:09:25.053 "num_base_bdevs_discovered": 3, 00:09:25.053 "num_base_bdevs_operational": 3, 00:09:25.053 "base_bdevs_list": [ 00:09:25.053 { 00:09:25.053 "name": "pt1", 00:09:25.053 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:25.053 "is_configured": true, 00:09:25.053 "data_offset": 2048, 00:09:25.053 "data_size": 63488 00:09:25.053 }, 00:09:25.053 { 00:09:25.053 "name": "pt2", 00:09:25.053 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:25.053 "is_configured": true, 00:09:25.053 "data_offset": 2048, 00:09:25.053 "data_size": 63488 00:09:25.053 }, 00:09:25.053 { 00:09:25.053 "name": "pt3", 00:09:25.053 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:25.053 "is_configured": true, 00:09:25.053 "data_offset": 2048, 00:09:25.053 "data_size": 63488 00:09:25.053 } 00:09:25.053 ] 00:09:25.053 }' 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:25.053 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.312 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:25.312 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:25.312 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:25.312 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:25.312 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:25.312 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:25.312 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:25.312 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:25.312 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.312 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.312 [2024-12-05 18:59:42.837781] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:25.312 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.571 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:25.571 "name": "raid_bdev1", 00:09:25.571 "aliases": [ 00:09:25.571 "0ed75771-455d-479f-ad9f-99c5bdb6d298" 00:09:25.571 ], 00:09:25.571 "product_name": "Raid Volume", 00:09:25.571 "block_size": 512, 00:09:25.571 "num_blocks": 63488, 00:09:25.571 "uuid": "0ed75771-455d-479f-ad9f-99c5bdb6d298", 00:09:25.571 "assigned_rate_limits": { 00:09:25.571 "rw_ios_per_sec": 0, 00:09:25.571 "rw_mbytes_per_sec": 0, 00:09:25.571 "r_mbytes_per_sec": 0, 00:09:25.571 "w_mbytes_per_sec": 0 00:09:25.571 }, 00:09:25.571 "claimed": false, 00:09:25.571 "zoned": false, 00:09:25.571 "supported_io_types": { 00:09:25.571 "read": true, 00:09:25.571 "write": true, 00:09:25.571 "unmap": false, 00:09:25.571 "flush": false, 00:09:25.571 "reset": true, 00:09:25.571 "nvme_admin": false, 00:09:25.571 "nvme_io": false, 00:09:25.571 "nvme_io_md": false, 00:09:25.571 "write_zeroes": true, 00:09:25.571 "zcopy": false, 00:09:25.571 "get_zone_info": false, 00:09:25.571 "zone_management": false, 00:09:25.571 "zone_append": false, 00:09:25.571 "compare": false, 00:09:25.571 "compare_and_write": false, 00:09:25.571 "abort": false, 00:09:25.571 "seek_hole": false, 00:09:25.571 "seek_data": false, 00:09:25.571 "copy": false, 00:09:25.571 "nvme_iov_md": false 00:09:25.571 }, 00:09:25.571 "memory_domains": [ 00:09:25.571 { 00:09:25.571 "dma_device_id": "system", 00:09:25.571 "dma_device_type": 1 00:09:25.571 }, 00:09:25.571 { 00:09:25.571 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.571 "dma_device_type": 2 00:09:25.571 }, 00:09:25.571 { 00:09:25.571 "dma_device_id": "system", 00:09:25.571 "dma_device_type": 1 00:09:25.571 }, 00:09:25.571 { 00:09:25.571 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.571 "dma_device_type": 2 00:09:25.571 }, 00:09:25.571 { 00:09:25.571 "dma_device_id": "system", 00:09:25.571 "dma_device_type": 1 00:09:25.571 }, 00:09:25.571 { 00:09:25.571 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.571 "dma_device_type": 2 00:09:25.571 } 00:09:25.571 ], 00:09:25.571 "driver_specific": { 00:09:25.571 "raid": { 00:09:25.571 "uuid": "0ed75771-455d-479f-ad9f-99c5bdb6d298", 00:09:25.571 "strip_size_kb": 0, 00:09:25.571 "state": "online", 00:09:25.571 "raid_level": "raid1", 00:09:25.571 "superblock": true, 00:09:25.571 "num_base_bdevs": 3, 00:09:25.571 "num_base_bdevs_discovered": 3, 00:09:25.571 "num_base_bdevs_operational": 3, 00:09:25.571 "base_bdevs_list": [ 00:09:25.571 { 00:09:25.571 "name": "pt1", 00:09:25.571 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:25.571 "is_configured": true, 00:09:25.571 "data_offset": 2048, 00:09:25.571 "data_size": 63488 00:09:25.571 }, 00:09:25.571 { 00:09:25.571 "name": "pt2", 00:09:25.571 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:25.571 "is_configured": true, 00:09:25.571 "data_offset": 2048, 00:09:25.571 "data_size": 63488 00:09:25.571 }, 00:09:25.571 { 00:09:25.571 "name": "pt3", 00:09:25.571 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:25.571 "is_configured": true, 00:09:25.571 "data_offset": 2048, 00:09:25.571 "data_size": 63488 00:09:25.571 } 00:09:25.571 ] 00:09:25.571 } 00:09:25.571 } 00:09:25.571 }' 00:09:25.571 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:25.571 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:25.571 pt2 00:09:25.571 pt3' 00:09:25.571 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:25.571 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:25.571 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:25.571 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:25.571 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.571 18:59:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:25.571 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.572 18:59:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.572 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:25.572 [2024-12-05 18:59:43.125237] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=0ed75771-455d-479f-ad9f-99c5bdb6d298 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 0ed75771-455d-479f-ad9f-99c5bdb6d298 ']' 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.832 [2024-12-05 18:59:43.172904] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:25.832 [2024-12-05 18:59:43.172926] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:25.832 [2024-12-05 18:59:43.172994] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:25.832 [2024-12-05 18:59:43.173067] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:25.832 [2024-12-05 18:59:43.173081] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:25.832 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.833 [2024-12-05 18:59:43.316717] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:25.833 [2024-12-05 18:59:43.318503] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:25.833 [2024-12-05 18:59:43.318541] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:25.833 [2024-12-05 18:59:43.318584] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:25.833 [2024-12-05 18:59:43.318634] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:25.833 [2024-12-05 18:59:43.318652] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:25.833 [2024-12-05 18:59:43.318718] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:25.833 [2024-12-05 18:59:43.318746] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:09:25.833 request: 00:09:25.833 { 00:09:25.833 "name": "raid_bdev1", 00:09:25.833 "raid_level": "raid1", 00:09:25.833 "base_bdevs": [ 00:09:25.833 "malloc1", 00:09:25.833 "malloc2", 00:09:25.833 "malloc3" 00:09:25.833 ], 00:09:25.833 "superblock": false, 00:09:25.833 "method": "bdev_raid_create", 00:09:25.833 "req_id": 1 00:09:25.833 } 00:09:25.833 Got JSON-RPC error response 00:09:25.833 response: 00:09:25.833 { 00:09:25.833 "code": -17, 00:09:25.833 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:25.833 } 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.833 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.833 [2024-12-05 18:59:43.388553] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:25.833 [2024-12-05 18:59:43.388603] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:25.833 [2024-12-05 18:59:43.388622] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:25.833 [2024-12-05 18:59:43.388633] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:26.093 [2024-12-05 18:59:43.390817] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:26.093 [2024-12-05 18:59:43.390855] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:26.093 [2024-12-05 18:59:43.390916] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:26.093 [2024-12-05 18:59:43.390962] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:26.093 pt1 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.093 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.093 "name": "raid_bdev1", 00:09:26.093 "uuid": "0ed75771-455d-479f-ad9f-99c5bdb6d298", 00:09:26.093 "strip_size_kb": 0, 00:09:26.093 "state": "configuring", 00:09:26.093 "raid_level": "raid1", 00:09:26.093 "superblock": true, 00:09:26.093 "num_base_bdevs": 3, 00:09:26.093 "num_base_bdevs_discovered": 1, 00:09:26.093 "num_base_bdevs_operational": 3, 00:09:26.093 "base_bdevs_list": [ 00:09:26.093 { 00:09:26.093 "name": "pt1", 00:09:26.093 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:26.093 "is_configured": true, 00:09:26.093 "data_offset": 2048, 00:09:26.093 "data_size": 63488 00:09:26.093 }, 00:09:26.093 { 00:09:26.093 "name": null, 00:09:26.093 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:26.093 "is_configured": false, 00:09:26.093 "data_offset": 2048, 00:09:26.093 "data_size": 63488 00:09:26.093 }, 00:09:26.093 { 00:09:26.093 "name": null, 00:09:26.093 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:26.093 "is_configured": false, 00:09:26.093 "data_offset": 2048, 00:09:26.094 "data_size": 63488 00:09:26.094 } 00:09:26.094 ] 00:09:26.094 }' 00:09:26.094 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.094 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.354 [2024-12-05 18:59:43.855754] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:26.354 [2024-12-05 18:59:43.855856] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:26.354 [2024-12-05 18:59:43.855879] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:09:26.354 [2024-12-05 18:59:43.855892] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:26.354 [2024-12-05 18:59:43.856237] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:26.354 [2024-12-05 18:59:43.856256] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:26.354 [2024-12-05 18:59:43.856314] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:26.354 [2024-12-05 18:59:43.856342] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:26.354 pt2 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.354 [2024-12-05 18:59:43.863763] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.354 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.614 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.614 "name": "raid_bdev1", 00:09:26.615 "uuid": "0ed75771-455d-479f-ad9f-99c5bdb6d298", 00:09:26.615 "strip_size_kb": 0, 00:09:26.615 "state": "configuring", 00:09:26.615 "raid_level": "raid1", 00:09:26.615 "superblock": true, 00:09:26.615 "num_base_bdevs": 3, 00:09:26.615 "num_base_bdevs_discovered": 1, 00:09:26.615 "num_base_bdevs_operational": 3, 00:09:26.615 "base_bdevs_list": [ 00:09:26.615 { 00:09:26.615 "name": "pt1", 00:09:26.615 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:26.615 "is_configured": true, 00:09:26.615 "data_offset": 2048, 00:09:26.615 "data_size": 63488 00:09:26.615 }, 00:09:26.615 { 00:09:26.615 "name": null, 00:09:26.615 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:26.615 "is_configured": false, 00:09:26.615 "data_offset": 0, 00:09:26.615 "data_size": 63488 00:09:26.615 }, 00:09:26.615 { 00:09:26.615 "name": null, 00:09:26.615 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:26.615 "is_configured": false, 00:09:26.615 "data_offset": 2048, 00:09:26.615 "data_size": 63488 00:09:26.615 } 00:09:26.615 ] 00:09:26.615 }' 00:09:26.615 18:59:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.615 18:59:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.916 [2024-12-05 18:59:44.334881] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:26.916 [2024-12-05 18:59:44.334973] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:26.916 [2024-12-05 18:59:44.335006] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:26.916 [2024-12-05 18:59:44.335033] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:26.916 [2024-12-05 18:59:44.335394] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:26.916 [2024-12-05 18:59:44.335449] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:26.916 [2024-12-05 18:59:44.335532] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:26.916 [2024-12-05 18:59:44.335575] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:26.916 pt2 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.916 [2024-12-05 18:59:44.346869] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:26.916 [2024-12-05 18:59:44.346948] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:26.916 [2024-12-05 18:59:44.346981] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:26.916 [2024-12-05 18:59:44.347006] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:26.916 [2024-12-05 18:59:44.347315] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:26.916 [2024-12-05 18:59:44.347368] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:26.916 [2024-12-05 18:59:44.347445] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:26.916 [2024-12-05 18:59:44.347485] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:26.916 [2024-12-05 18:59:44.347601] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:26.916 [2024-12-05 18:59:44.347636] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:26.916 [2024-12-05 18:59:44.347899] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:26.916 [2024-12-05 18:59:44.348047] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:26.916 [2024-12-05 18:59:44.348087] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:09:26.916 [2024-12-05 18:59:44.348215] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:26.916 pt3 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:26.916 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.916 "name": "raid_bdev1", 00:09:26.916 "uuid": "0ed75771-455d-479f-ad9f-99c5bdb6d298", 00:09:26.916 "strip_size_kb": 0, 00:09:26.916 "state": "online", 00:09:26.916 "raid_level": "raid1", 00:09:26.916 "superblock": true, 00:09:26.916 "num_base_bdevs": 3, 00:09:26.916 "num_base_bdevs_discovered": 3, 00:09:26.917 "num_base_bdevs_operational": 3, 00:09:26.917 "base_bdevs_list": [ 00:09:26.917 { 00:09:26.917 "name": "pt1", 00:09:26.917 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:26.917 "is_configured": true, 00:09:26.917 "data_offset": 2048, 00:09:26.917 "data_size": 63488 00:09:26.917 }, 00:09:26.917 { 00:09:26.917 "name": "pt2", 00:09:26.917 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:26.917 "is_configured": true, 00:09:26.917 "data_offset": 2048, 00:09:26.917 "data_size": 63488 00:09:26.917 }, 00:09:26.917 { 00:09:26.917 "name": "pt3", 00:09:26.917 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:26.917 "is_configured": true, 00:09:26.917 "data_offset": 2048, 00:09:26.917 "data_size": 63488 00:09:26.917 } 00:09:26.917 ] 00:09:26.917 }' 00:09:26.917 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.917 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.487 [2024-12-05 18:59:44.762453] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:27.487 "name": "raid_bdev1", 00:09:27.487 "aliases": [ 00:09:27.487 "0ed75771-455d-479f-ad9f-99c5bdb6d298" 00:09:27.487 ], 00:09:27.487 "product_name": "Raid Volume", 00:09:27.487 "block_size": 512, 00:09:27.487 "num_blocks": 63488, 00:09:27.487 "uuid": "0ed75771-455d-479f-ad9f-99c5bdb6d298", 00:09:27.487 "assigned_rate_limits": { 00:09:27.487 "rw_ios_per_sec": 0, 00:09:27.487 "rw_mbytes_per_sec": 0, 00:09:27.487 "r_mbytes_per_sec": 0, 00:09:27.487 "w_mbytes_per_sec": 0 00:09:27.487 }, 00:09:27.487 "claimed": false, 00:09:27.487 "zoned": false, 00:09:27.487 "supported_io_types": { 00:09:27.487 "read": true, 00:09:27.487 "write": true, 00:09:27.487 "unmap": false, 00:09:27.487 "flush": false, 00:09:27.487 "reset": true, 00:09:27.487 "nvme_admin": false, 00:09:27.487 "nvme_io": false, 00:09:27.487 "nvme_io_md": false, 00:09:27.487 "write_zeroes": true, 00:09:27.487 "zcopy": false, 00:09:27.487 "get_zone_info": false, 00:09:27.487 "zone_management": false, 00:09:27.487 "zone_append": false, 00:09:27.487 "compare": false, 00:09:27.487 "compare_and_write": false, 00:09:27.487 "abort": false, 00:09:27.487 "seek_hole": false, 00:09:27.487 "seek_data": false, 00:09:27.487 "copy": false, 00:09:27.487 "nvme_iov_md": false 00:09:27.487 }, 00:09:27.487 "memory_domains": [ 00:09:27.487 { 00:09:27.487 "dma_device_id": "system", 00:09:27.487 "dma_device_type": 1 00:09:27.487 }, 00:09:27.487 { 00:09:27.487 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:27.487 "dma_device_type": 2 00:09:27.487 }, 00:09:27.487 { 00:09:27.487 "dma_device_id": "system", 00:09:27.487 "dma_device_type": 1 00:09:27.487 }, 00:09:27.487 { 00:09:27.487 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:27.487 "dma_device_type": 2 00:09:27.487 }, 00:09:27.487 { 00:09:27.487 "dma_device_id": "system", 00:09:27.487 "dma_device_type": 1 00:09:27.487 }, 00:09:27.487 { 00:09:27.487 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:27.487 "dma_device_type": 2 00:09:27.487 } 00:09:27.487 ], 00:09:27.487 "driver_specific": { 00:09:27.487 "raid": { 00:09:27.487 "uuid": "0ed75771-455d-479f-ad9f-99c5bdb6d298", 00:09:27.487 "strip_size_kb": 0, 00:09:27.487 "state": "online", 00:09:27.487 "raid_level": "raid1", 00:09:27.487 "superblock": true, 00:09:27.487 "num_base_bdevs": 3, 00:09:27.487 "num_base_bdevs_discovered": 3, 00:09:27.487 "num_base_bdevs_operational": 3, 00:09:27.487 "base_bdevs_list": [ 00:09:27.487 { 00:09:27.487 "name": "pt1", 00:09:27.487 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:27.487 "is_configured": true, 00:09:27.487 "data_offset": 2048, 00:09:27.487 "data_size": 63488 00:09:27.487 }, 00:09:27.487 { 00:09:27.487 "name": "pt2", 00:09:27.487 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:27.487 "is_configured": true, 00:09:27.487 "data_offset": 2048, 00:09:27.487 "data_size": 63488 00:09:27.487 }, 00:09:27.487 { 00:09:27.487 "name": "pt3", 00:09:27.487 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:27.487 "is_configured": true, 00:09:27.487 "data_offset": 2048, 00:09:27.487 "data_size": 63488 00:09:27.487 } 00:09:27.487 ] 00:09:27.487 } 00:09:27.487 } 00:09:27.487 }' 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:27.487 pt2 00:09:27.487 pt3' 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:27.487 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.488 18:59:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.488 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.488 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:27.488 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:27.488 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:27.488 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.488 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.488 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:27.488 [2024-12-05 18:59:45.033951] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:27.760 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.760 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 0ed75771-455d-479f-ad9f-99c5bdb6d298 '!=' 0ed75771-455d-479f-ad9f-99c5bdb6d298 ']' 00:09:27.760 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:09:27.760 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:27.760 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:27.760 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:09:27.760 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.760 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.760 [2024-12-05 18:59:45.061695] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:09:27.760 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.760 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:27.760 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:27.760 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:27.760 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:27.760 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:27.760 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:27.761 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.761 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.761 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.761 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.761 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:27.761 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.761 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:27.761 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.761 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:27.761 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.761 "name": "raid_bdev1", 00:09:27.761 "uuid": "0ed75771-455d-479f-ad9f-99c5bdb6d298", 00:09:27.761 "strip_size_kb": 0, 00:09:27.761 "state": "online", 00:09:27.761 "raid_level": "raid1", 00:09:27.761 "superblock": true, 00:09:27.761 "num_base_bdevs": 3, 00:09:27.761 "num_base_bdevs_discovered": 2, 00:09:27.761 "num_base_bdevs_operational": 2, 00:09:27.761 "base_bdevs_list": [ 00:09:27.761 { 00:09:27.761 "name": null, 00:09:27.761 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.761 "is_configured": false, 00:09:27.761 "data_offset": 0, 00:09:27.761 "data_size": 63488 00:09:27.761 }, 00:09:27.761 { 00:09:27.761 "name": "pt2", 00:09:27.761 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:27.761 "is_configured": true, 00:09:27.761 "data_offset": 2048, 00:09:27.761 "data_size": 63488 00:09:27.761 }, 00:09:27.761 { 00:09:27.761 "name": "pt3", 00:09:27.761 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:27.761 "is_configured": true, 00:09:27.761 "data_offset": 2048, 00:09:27.762 "data_size": 63488 00:09:27.762 } 00:09:27.762 ] 00:09:27.762 }' 00:09:27.762 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.762 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.024 [2024-12-05 18:59:45.504914] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:28.024 [2024-12-05 18:59:45.504987] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:28.024 [2024-12-05 18:59:45.505074] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:28.024 [2024-12-05 18:59:45.505142] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:28.024 [2024-12-05 18:59:45.505184] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.024 [2024-12-05 18:59:45.572784] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:28.024 [2024-12-05 18:59:45.572833] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:28.024 [2024-12-05 18:59:45.572849] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:09:28.024 [2024-12-05 18:59:45.572858] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:28.024 [2024-12-05 18:59:45.574960] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:28.024 [2024-12-05 18:59:45.574996] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:28.024 [2024-12-05 18:59:45.575060] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:28.024 [2024-12-05 18:59:45.575090] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:28.024 pt2 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:28.024 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:28.283 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.283 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:28.283 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.283 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.283 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.283 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:28.283 "name": "raid_bdev1", 00:09:28.283 "uuid": "0ed75771-455d-479f-ad9f-99c5bdb6d298", 00:09:28.283 "strip_size_kb": 0, 00:09:28.283 "state": "configuring", 00:09:28.283 "raid_level": "raid1", 00:09:28.283 "superblock": true, 00:09:28.283 "num_base_bdevs": 3, 00:09:28.283 "num_base_bdevs_discovered": 1, 00:09:28.283 "num_base_bdevs_operational": 2, 00:09:28.283 "base_bdevs_list": [ 00:09:28.283 { 00:09:28.283 "name": null, 00:09:28.283 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.283 "is_configured": false, 00:09:28.283 "data_offset": 2048, 00:09:28.283 "data_size": 63488 00:09:28.283 }, 00:09:28.283 { 00:09:28.283 "name": "pt2", 00:09:28.283 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:28.283 "is_configured": true, 00:09:28.283 "data_offset": 2048, 00:09:28.283 "data_size": 63488 00:09:28.283 }, 00:09:28.283 { 00:09:28.283 "name": null, 00:09:28.283 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:28.283 "is_configured": false, 00:09:28.283 "data_offset": 2048, 00:09:28.283 "data_size": 63488 00:09:28.283 } 00:09:28.283 ] 00:09:28.283 }' 00:09:28.283 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:28.283 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.543 [2024-12-05 18:59:45.956142] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:28.543 [2024-12-05 18:59:45.956243] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:28.543 [2024-12-05 18:59:45.956284] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:09:28.543 [2024-12-05 18:59:45.956312] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:28.543 [2024-12-05 18:59:45.956680] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:28.543 [2024-12-05 18:59:45.956733] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:28.543 [2024-12-05 18:59:45.956821] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:28.543 [2024-12-05 18:59:45.956868] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:28.543 [2024-12-05 18:59:45.956970] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:28.543 [2024-12-05 18:59:45.957005] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:28.543 [2024-12-05 18:59:45.957257] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:28.543 [2024-12-05 18:59:45.957411] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:28.543 [2024-12-05 18:59:45.957456] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:28.543 [2024-12-05 18:59:45.957587] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:28.543 pt3 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.543 18:59:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:28.543 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:28.543 "name": "raid_bdev1", 00:09:28.543 "uuid": "0ed75771-455d-479f-ad9f-99c5bdb6d298", 00:09:28.543 "strip_size_kb": 0, 00:09:28.543 "state": "online", 00:09:28.543 "raid_level": "raid1", 00:09:28.543 "superblock": true, 00:09:28.543 "num_base_bdevs": 3, 00:09:28.543 "num_base_bdevs_discovered": 2, 00:09:28.543 "num_base_bdevs_operational": 2, 00:09:28.543 "base_bdevs_list": [ 00:09:28.543 { 00:09:28.543 "name": null, 00:09:28.543 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.543 "is_configured": false, 00:09:28.543 "data_offset": 2048, 00:09:28.543 "data_size": 63488 00:09:28.543 }, 00:09:28.543 { 00:09:28.543 "name": "pt2", 00:09:28.543 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:28.543 "is_configured": true, 00:09:28.543 "data_offset": 2048, 00:09:28.543 "data_size": 63488 00:09:28.543 }, 00:09:28.543 { 00:09:28.543 "name": "pt3", 00:09:28.543 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:28.543 "is_configured": true, 00:09:28.543 "data_offset": 2048, 00:09:28.543 "data_size": 63488 00:09:28.543 } 00:09:28.543 ] 00:09:28.543 }' 00:09:28.543 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:28.543 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.114 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:29.114 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.114 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.114 [2024-12-05 18:59:46.379456] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:29.114 [2024-12-05 18:59:46.379526] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:29.114 [2024-12-05 18:59:46.379604] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:29.114 [2024-12-05 18:59:46.379679] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:29.114 [2024-12-05 18:59:46.379739] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:29.114 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.114 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:09:29.114 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.114 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.114 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.114 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.115 [2024-12-05 18:59:46.431364] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:29.115 [2024-12-05 18:59:46.431414] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:29.115 [2024-12-05 18:59:46.431448] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:29.115 [2024-12-05 18:59:46.431458] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:29.115 [2024-12-05 18:59:46.433489] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:29.115 [2024-12-05 18:59:46.433527] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:29.115 [2024-12-05 18:59:46.433584] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:29.115 [2024-12-05 18:59:46.433626] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:29.115 [2024-12-05 18:59:46.433741] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:09:29.115 [2024-12-05 18:59:46.433757] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:29.115 [2024-12-05 18:59:46.433778] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:09:29.115 [2024-12-05 18:59:46.433822] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:29.115 pt1 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.115 "name": "raid_bdev1", 00:09:29.115 "uuid": "0ed75771-455d-479f-ad9f-99c5bdb6d298", 00:09:29.115 "strip_size_kb": 0, 00:09:29.115 "state": "configuring", 00:09:29.115 "raid_level": "raid1", 00:09:29.115 "superblock": true, 00:09:29.115 "num_base_bdevs": 3, 00:09:29.115 "num_base_bdevs_discovered": 1, 00:09:29.115 "num_base_bdevs_operational": 2, 00:09:29.115 "base_bdevs_list": [ 00:09:29.115 { 00:09:29.115 "name": null, 00:09:29.115 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:29.115 "is_configured": false, 00:09:29.115 "data_offset": 2048, 00:09:29.115 "data_size": 63488 00:09:29.115 }, 00:09:29.115 { 00:09:29.115 "name": "pt2", 00:09:29.115 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:29.115 "is_configured": true, 00:09:29.115 "data_offset": 2048, 00:09:29.115 "data_size": 63488 00:09:29.115 }, 00:09:29.115 { 00:09:29.115 "name": null, 00:09:29.115 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:29.115 "is_configured": false, 00:09:29.115 "data_offset": 2048, 00:09:29.115 "data_size": 63488 00:09:29.115 } 00:09:29.115 ] 00:09:29.115 }' 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.115 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.376 [2024-12-05 18:59:46.910554] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:29.376 [2024-12-05 18:59:46.910614] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:29.376 [2024-12-05 18:59:46.910631] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:09:29.376 [2024-12-05 18:59:46.910641] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:29.376 [2024-12-05 18:59:46.911001] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:29.376 [2024-12-05 18:59:46.911024] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:29.376 [2024-12-05 18:59:46.911086] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:29.376 [2024-12-05 18:59:46.911110] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:29.376 [2024-12-05 18:59:46.911188] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:09:29.376 [2024-12-05 18:59:46.911199] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:29.376 [2024-12-05 18:59:46.911415] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:09:29.376 [2024-12-05 18:59:46.911534] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:09:29.376 [2024-12-05 18:59:46.911541] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:09:29.376 [2024-12-05 18:59:46.911638] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:29.376 pt3 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:29.376 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.637 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.637 "name": "raid_bdev1", 00:09:29.637 "uuid": "0ed75771-455d-479f-ad9f-99c5bdb6d298", 00:09:29.637 "strip_size_kb": 0, 00:09:29.637 "state": "online", 00:09:29.637 "raid_level": "raid1", 00:09:29.637 "superblock": true, 00:09:29.637 "num_base_bdevs": 3, 00:09:29.637 "num_base_bdevs_discovered": 2, 00:09:29.637 "num_base_bdevs_operational": 2, 00:09:29.637 "base_bdevs_list": [ 00:09:29.637 { 00:09:29.637 "name": null, 00:09:29.637 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:29.637 "is_configured": false, 00:09:29.637 "data_offset": 2048, 00:09:29.637 "data_size": 63488 00:09:29.637 }, 00:09:29.637 { 00:09:29.637 "name": "pt2", 00:09:29.637 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:29.637 "is_configured": true, 00:09:29.637 "data_offset": 2048, 00:09:29.637 "data_size": 63488 00:09:29.637 }, 00:09:29.637 { 00:09:29.637 "name": "pt3", 00:09:29.637 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:29.637 "is_configured": true, 00:09:29.637 "data_offset": 2048, 00:09:29.637 "data_size": 63488 00:09:29.637 } 00:09:29.637 ] 00:09:29.637 }' 00:09:29.637 18:59:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.637 18:59:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.897 18:59:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:29.897 18:59:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:09:29.897 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.897 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.897 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.897 18:59:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:09:29.897 18:59:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:29.897 18:59:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:09:29.897 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:29.897 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.897 [2024-12-05 18:59:47.417926] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:29.897 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:29.897 18:59:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 0ed75771-455d-479f-ad9f-99c5bdb6d298 '!=' 0ed75771-455d-479f-ad9f-99c5bdb6d298 ']' 00:09:29.897 18:59:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 79066 00:09:29.897 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 79066 ']' 00:09:29.897 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 79066 00:09:29.898 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:09:29.898 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:29.898 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79066 00:09:30.158 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:30.158 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:30.158 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79066' 00:09:30.158 killing process with pid 79066 00:09:30.158 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 79066 00:09:30.158 [2024-12-05 18:59:47.478022] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:30.158 [2024-12-05 18:59:47.478132] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:30.158 [2024-12-05 18:59:47.478215] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to fr 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 79066 00:09:30.158 ee all in destruct 00:09:30.158 [2024-12-05 18:59:47.478275] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:09:30.158 [2024-12-05 18:59:47.510709] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:30.419 18:59:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:30.419 00:09:30.419 real 0m6.335s 00:09:30.419 user 0m10.703s 00:09:30.419 sys 0m1.235s 00:09:30.419 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:30.419 18:59:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.419 ************************************ 00:09:30.419 END TEST raid_superblock_test 00:09:30.419 ************************************ 00:09:30.419 18:59:47 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 3 read 00:09:30.419 18:59:47 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:30.419 18:59:47 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:30.419 18:59:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:30.419 ************************************ 00:09:30.419 START TEST raid_read_error_test 00:09:30.419 ************************************ 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 3 read 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.9HFbfAV8wp 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=79459 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 79459 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 79459 ']' 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:30.419 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:30.419 18:59:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.419 [2024-12-05 18:59:47.898262] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:09:30.419 [2024-12-05 18:59:47.898373] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79459 ] 00:09:30.679 [2024-12-05 18:59:48.052908] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:30.679 [2024-12-05 18:59:48.077683] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:30.680 [2024-12-05 18:59:48.120705] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:30.680 [2024-12-05 18:59:48.120741] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.251 BaseBdev1_malloc 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.251 true 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.251 [2024-12-05 18:59:48.745051] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:31.251 [2024-12-05 18:59:48.745099] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:31.251 [2024-12-05 18:59:48.745120] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:31.251 [2024-12-05 18:59:48.745128] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:31.251 [2024-12-05 18:59:48.747327] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:31.251 [2024-12-05 18:59:48.747428] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:31.251 BaseBdev1 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.251 BaseBdev2_malloc 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.251 true 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.251 [2024-12-05 18:59:48.785877] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:31.251 [2024-12-05 18:59:48.785984] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:31.251 [2024-12-05 18:59:48.786005] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:31.251 [2024-12-05 18:59:48.786022] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:31.251 [2024-12-05 18:59:48.788136] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:31.251 [2024-12-05 18:59:48.788177] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:31.251 BaseBdev2 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.251 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.512 BaseBdev3_malloc 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.512 true 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.512 [2024-12-05 18:59:48.826717] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:31.512 [2024-12-05 18:59:48.826820] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:31.512 [2024-12-05 18:59:48.826842] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:31.512 [2024-12-05 18:59:48.826850] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:31.512 [2024-12-05 18:59:48.828921] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:31.512 [2024-12-05 18:59:48.828958] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:31.512 BaseBdev3 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.512 [2024-12-05 18:59:48.838758] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:31.512 [2024-12-05 18:59:48.840568] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:31.512 [2024-12-05 18:59:48.840641] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:31.512 [2024-12-05 18:59:48.840826] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:31.512 [2024-12-05 18:59:48.840842] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:31.512 [2024-12-05 18:59:48.841089] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:09:31.512 [2024-12-05 18:59:48.841233] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:31.512 [2024-12-05 18:59:48.841242] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:31.512 [2024-12-05 18:59:48.841384] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.512 "name": "raid_bdev1", 00:09:31.512 "uuid": "0feab66e-70d4-45f2-93c4-f8fb037b6ada", 00:09:31.512 "strip_size_kb": 0, 00:09:31.512 "state": "online", 00:09:31.512 "raid_level": "raid1", 00:09:31.512 "superblock": true, 00:09:31.512 "num_base_bdevs": 3, 00:09:31.512 "num_base_bdevs_discovered": 3, 00:09:31.512 "num_base_bdevs_operational": 3, 00:09:31.512 "base_bdevs_list": [ 00:09:31.512 { 00:09:31.512 "name": "BaseBdev1", 00:09:31.512 "uuid": "0984ace4-338e-54f5-bf2d-f586ea65ab83", 00:09:31.512 "is_configured": true, 00:09:31.512 "data_offset": 2048, 00:09:31.512 "data_size": 63488 00:09:31.512 }, 00:09:31.512 { 00:09:31.512 "name": "BaseBdev2", 00:09:31.512 "uuid": "c1fea918-2384-59e5-a8e5-f1d072628121", 00:09:31.512 "is_configured": true, 00:09:31.512 "data_offset": 2048, 00:09:31.512 "data_size": 63488 00:09:31.512 }, 00:09:31.512 { 00:09:31.512 "name": "BaseBdev3", 00:09:31.512 "uuid": "321d0945-7fbb-5f6a-a9a6-ec25264ace65", 00:09:31.512 "is_configured": true, 00:09:31.512 "data_offset": 2048, 00:09:31.512 "data_size": 63488 00:09:31.512 } 00:09:31.512 ] 00:09:31.512 }' 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.512 18:59:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.791 18:59:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:31.791 18:59:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:31.791 [2024-12-05 18:59:49.330267] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:32.731 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.992 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:32.992 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:32.992 "name": "raid_bdev1", 00:09:32.992 "uuid": "0feab66e-70d4-45f2-93c4-f8fb037b6ada", 00:09:32.992 "strip_size_kb": 0, 00:09:32.992 "state": "online", 00:09:32.992 "raid_level": "raid1", 00:09:32.992 "superblock": true, 00:09:32.992 "num_base_bdevs": 3, 00:09:32.992 "num_base_bdevs_discovered": 3, 00:09:32.992 "num_base_bdevs_operational": 3, 00:09:32.992 "base_bdevs_list": [ 00:09:32.992 { 00:09:32.992 "name": "BaseBdev1", 00:09:32.992 "uuid": "0984ace4-338e-54f5-bf2d-f586ea65ab83", 00:09:32.992 "is_configured": true, 00:09:32.992 "data_offset": 2048, 00:09:32.992 "data_size": 63488 00:09:32.992 }, 00:09:32.992 { 00:09:32.992 "name": "BaseBdev2", 00:09:32.992 "uuid": "c1fea918-2384-59e5-a8e5-f1d072628121", 00:09:32.992 "is_configured": true, 00:09:32.992 "data_offset": 2048, 00:09:32.992 "data_size": 63488 00:09:32.992 }, 00:09:32.992 { 00:09:32.992 "name": "BaseBdev3", 00:09:32.992 "uuid": "321d0945-7fbb-5f6a-a9a6-ec25264ace65", 00:09:32.992 "is_configured": true, 00:09:32.992 "data_offset": 2048, 00:09:32.992 "data_size": 63488 00:09:32.992 } 00:09:32.992 ] 00:09:32.992 }' 00:09:32.992 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:32.992 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.252 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:33.252 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:33.252 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.252 [2024-12-05 18:59:50.725147] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:33.252 [2024-12-05 18:59:50.725251] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:33.252 [2024-12-05 18:59:50.727893] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:33.252 [2024-12-05 18:59:50.727987] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:33.252 [2024-12-05 18:59:50.728107] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:33.252 [2024-12-05 18:59:50.728171] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:33.252 { 00:09:33.252 "results": [ 00:09:33.252 { 00:09:33.252 "job": "raid_bdev1", 00:09:33.252 "core_mask": "0x1", 00:09:33.252 "workload": "randrw", 00:09:33.252 "percentage": 50, 00:09:33.252 "status": "finished", 00:09:33.252 "queue_depth": 1, 00:09:33.252 "io_size": 131072, 00:09:33.252 "runtime": 1.396008, 00:09:33.252 "iops": 14795.76048274795, 00:09:33.252 "mibps": 1849.4700603434937, 00:09:33.252 "io_failed": 0, 00:09:33.252 "io_timeout": 0, 00:09:33.252 "avg_latency_us": 65.09697198411415, 00:09:33.252 "min_latency_us": 22.46986899563319, 00:09:33.252 "max_latency_us": 1402.2986899563318 00:09:33.252 } 00:09:33.252 ], 00:09:33.252 "core_count": 1 00:09:33.252 } 00:09:33.252 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:33.252 18:59:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 79459 00:09:33.252 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 79459 ']' 00:09:33.252 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 79459 00:09:33.252 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:09:33.252 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:33.252 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79459 00:09:33.252 killing process with pid 79459 00:09:33.252 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:33.252 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:33.252 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79459' 00:09:33.252 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 79459 00:09:33.252 [2024-12-05 18:59:50.772387] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:33.252 18:59:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 79459 00:09:33.252 [2024-12-05 18:59:50.797448] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:33.512 18:59:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.9HFbfAV8wp 00:09:33.512 18:59:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:33.512 18:59:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:33.512 18:59:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:33.512 18:59:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:33.512 ************************************ 00:09:33.512 END TEST raid_read_error_test 00:09:33.512 ************************************ 00:09:33.512 18:59:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:33.512 18:59:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:33.512 18:59:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:33.512 00:09:33.512 real 0m3.213s 00:09:33.512 user 0m4.067s 00:09:33.512 sys 0m0.495s 00:09:33.512 18:59:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:33.512 18:59:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.771 18:59:51 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 3 write 00:09:33.771 18:59:51 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:33.771 18:59:51 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:33.771 18:59:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:33.771 ************************************ 00:09:33.771 START TEST raid_write_error_test 00:09:33.771 ************************************ 00:09:33.771 18:59:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 3 write 00:09:33.771 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:33.771 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:33.771 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:33.771 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:33.771 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:33.771 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:33.771 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:33.771 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:33.771 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:33.771 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.MUlf7UPFiT 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=79570 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 79570 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 79570 ']' 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:33.772 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:33.772 18:59:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.772 [2024-12-05 18:59:51.184751] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:09:33.772 [2024-12-05 18:59:51.184943] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79570 ] 00:09:34.031 [2024-12-05 18:59:51.337194] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:34.032 [2024-12-05 18:59:51.361909] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:34.032 [2024-12-05 18:59:51.404962] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:34.032 [2024-12-05 18:59:51.405007] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.603 BaseBdev1_malloc 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.603 true 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.603 [2024-12-05 18:59:52.033083] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:34.603 [2024-12-05 18:59:52.033185] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:34.603 [2024-12-05 18:59:52.033241] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:34.603 [2024-12-05 18:59:52.033276] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:34.603 [2024-12-05 18:59:52.035359] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:34.603 [2024-12-05 18:59:52.035446] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:34.603 BaseBdev1 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.603 BaseBdev2_malloc 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.603 true 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.603 [2024-12-05 18:59:52.061737] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:34.603 [2024-12-05 18:59:52.061834] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:34.603 [2024-12-05 18:59:52.061867] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:34.603 [2024-12-05 18:59:52.061903] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:34.603 [2024-12-05 18:59:52.064009] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:34.603 [2024-12-05 18:59:52.064084] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:34.603 BaseBdev2 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.603 BaseBdev3_malloc 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.603 true 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.603 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.603 [2024-12-05 18:59:52.102334] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:34.603 [2024-12-05 18:59:52.102377] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:34.603 [2024-12-05 18:59:52.102410] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:34.603 [2024-12-05 18:59:52.102418] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:34.604 [2024-12-05 18:59:52.104506] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:34.604 [2024-12-05 18:59:52.104544] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:34.604 BaseBdev3 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.604 [2024-12-05 18:59:52.114368] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:34.604 [2024-12-05 18:59:52.116268] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:34.604 [2024-12-05 18:59:52.116339] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:34.604 [2024-12-05 18:59:52.116505] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:34.604 [2024-12-05 18:59:52.116520] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:34.604 [2024-12-05 18:59:52.116769] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002bb0 00:09:34.604 [2024-12-05 18:59:52.116924] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:34.604 [2024-12-05 18:59:52.116961] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:34.604 [2024-12-05 18:59:52.117081] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:34.604 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:34.864 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:34.864 "name": "raid_bdev1", 00:09:34.864 "uuid": "3ed90629-1659-4af7-9737-ba7134b33bfd", 00:09:34.864 "strip_size_kb": 0, 00:09:34.864 "state": "online", 00:09:34.864 "raid_level": "raid1", 00:09:34.864 "superblock": true, 00:09:34.864 "num_base_bdevs": 3, 00:09:34.864 "num_base_bdevs_discovered": 3, 00:09:34.864 "num_base_bdevs_operational": 3, 00:09:34.864 "base_bdevs_list": [ 00:09:34.864 { 00:09:34.864 "name": "BaseBdev1", 00:09:34.864 "uuid": "2ce414a2-ab28-569d-b5f5-601f081e8180", 00:09:34.864 "is_configured": true, 00:09:34.864 "data_offset": 2048, 00:09:34.864 "data_size": 63488 00:09:34.864 }, 00:09:34.864 { 00:09:34.864 "name": "BaseBdev2", 00:09:34.864 "uuid": "9e48c2dc-6009-59f8-a160-470270df04f3", 00:09:34.864 "is_configured": true, 00:09:34.864 "data_offset": 2048, 00:09:34.864 "data_size": 63488 00:09:34.864 }, 00:09:34.864 { 00:09:34.864 "name": "BaseBdev3", 00:09:34.864 "uuid": "90bc26b4-acfa-55ca-a0e9-98dde6e5ae4e", 00:09:34.864 "is_configured": true, 00:09:34.864 "data_offset": 2048, 00:09:34.864 "data_size": 63488 00:09:34.864 } 00:09:34.864 ] 00:09:34.864 }' 00:09:34.864 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:34.864 18:59:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:35.124 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:35.124 18:59:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:35.124 [2024-12-05 18:59:52.601920] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002d50 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.065 [2024-12-05 18:59:53.528722] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:09:36.065 [2024-12-05 18:59:53.528778] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:36.065 [2024-12-05 18:59:53.529009] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002d50 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.065 "name": "raid_bdev1", 00:09:36.065 "uuid": "3ed90629-1659-4af7-9737-ba7134b33bfd", 00:09:36.065 "strip_size_kb": 0, 00:09:36.065 "state": "online", 00:09:36.065 "raid_level": "raid1", 00:09:36.065 "superblock": true, 00:09:36.065 "num_base_bdevs": 3, 00:09:36.065 "num_base_bdevs_discovered": 2, 00:09:36.065 "num_base_bdevs_operational": 2, 00:09:36.065 "base_bdevs_list": [ 00:09:36.065 { 00:09:36.065 "name": null, 00:09:36.065 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:36.065 "is_configured": false, 00:09:36.065 "data_offset": 0, 00:09:36.065 "data_size": 63488 00:09:36.065 }, 00:09:36.065 { 00:09:36.065 "name": "BaseBdev2", 00:09:36.065 "uuid": "9e48c2dc-6009-59f8-a160-470270df04f3", 00:09:36.065 "is_configured": true, 00:09:36.065 "data_offset": 2048, 00:09:36.065 "data_size": 63488 00:09:36.065 }, 00:09:36.065 { 00:09:36.065 "name": "BaseBdev3", 00:09:36.065 "uuid": "90bc26b4-acfa-55ca-a0e9-98dde6e5ae4e", 00:09:36.065 "is_configured": true, 00:09:36.065 "data_offset": 2048, 00:09:36.065 "data_size": 63488 00:09:36.065 } 00:09:36.065 ] 00:09:36.065 }' 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.065 18:59:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.635 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:36.635 18:59:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:36.635 18:59:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.635 [2024-12-05 18:59:53.978791] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:36.635 [2024-12-05 18:59:53.978893] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:36.635 [2024-12-05 18:59:53.981381] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:36.635 [2024-12-05 18:59:53.981468] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:36.635 [2024-12-05 18:59:53.981567] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:36.635 [2024-12-05 18:59:53.981622] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:36.635 { 00:09:36.635 "results": [ 00:09:36.635 { 00:09:36.635 "job": "raid_bdev1", 00:09:36.635 "core_mask": "0x1", 00:09:36.635 "workload": "randrw", 00:09:36.635 "percentage": 50, 00:09:36.635 "status": "finished", 00:09:36.635 "queue_depth": 1, 00:09:36.635 "io_size": 131072, 00:09:36.635 "runtime": 1.377717, 00:09:36.635 "iops": 16525.890295321897, 00:09:36.635 "mibps": 2065.736286915237, 00:09:36.635 "io_failed": 0, 00:09:36.635 "io_timeout": 0, 00:09:36.635 "avg_latency_us": 57.972716476353845, 00:09:36.635 "min_latency_us": 22.46986899563319, 00:09:36.635 "max_latency_us": 1380.8349344978167 00:09:36.635 } 00:09:36.635 ], 00:09:36.635 "core_count": 1 00:09:36.635 } 00:09:36.635 18:59:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:36.635 18:59:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 79570 00:09:36.635 18:59:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 79570 ']' 00:09:36.635 18:59:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 79570 00:09:36.635 18:59:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:09:36.635 18:59:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:36.635 18:59:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79570 00:09:36.635 18:59:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:36.635 18:59:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:36.635 18:59:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79570' 00:09:36.635 killing process with pid 79570 00:09:36.635 18:59:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 79570 00:09:36.635 [2024-12-05 18:59:54.028266] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:36.635 18:59:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 79570 00:09:36.635 [2024-12-05 18:59:54.054126] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:36.895 18:59:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:36.895 18:59:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.MUlf7UPFiT 00:09:36.895 18:59:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:36.895 18:59:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:36.895 18:59:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:36.895 ************************************ 00:09:36.895 END TEST raid_write_error_test 00:09:36.895 ************************************ 00:09:36.895 18:59:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:36.895 18:59:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:36.895 18:59:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:36.895 00:09:36.895 real 0m3.181s 00:09:36.895 user 0m4.003s 00:09:36.895 sys 0m0.523s 00:09:36.895 18:59:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:36.895 18:59:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.895 18:59:54 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:09:36.895 18:59:54 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:36.895 18:59:54 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 4 false 00:09:36.895 18:59:54 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:36.895 18:59:54 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:36.895 18:59:54 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:36.895 ************************************ 00:09:36.895 START TEST raid_state_function_test 00:09:36.895 ************************************ 00:09:36.895 18:59:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 4 false 00:09:36.895 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:36.895 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:36.895 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:36.895 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:36.895 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:36.895 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:36.895 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:36.895 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:36.895 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:36.895 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:36.895 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:36.895 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:36.895 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:36.895 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:36.896 Process raid pid: 79679 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=79679 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 79679' 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 79679 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 79679 ']' 00:09:36.896 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:36.896 18:59:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:36.896 [2024-12-05 18:59:54.435116] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:09:36.896 [2024-12-05 18:59:54.435257] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:37.156 [2024-12-05 18:59:54.592226] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:37.156 [2024-12-05 18:59:54.617204] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:37.156 [2024-12-05 18:59:54.660156] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:37.156 [2024-12-05 18:59:54.660194] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.727 [2024-12-05 18:59:55.255267] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:37.727 [2024-12-05 18:59:55.255334] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:37.727 [2024-12-05 18:59:55.255351] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:37.727 [2024-12-05 18:59:55.255364] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:37.727 [2024-12-05 18:59:55.255370] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:37.727 [2024-12-05 18:59:55.255381] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:37.727 [2024-12-05 18:59:55.255387] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:37.727 [2024-12-05 18:59:55.255396] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:37.727 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:37.988 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:37.988 "name": "Existed_Raid", 00:09:37.988 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:37.988 "strip_size_kb": 64, 00:09:37.988 "state": "configuring", 00:09:37.988 "raid_level": "raid0", 00:09:37.988 "superblock": false, 00:09:37.988 "num_base_bdevs": 4, 00:09:37.988 "num_base_bdevs_discovered": 0, 00:09:37.988 "num_base_bdevs_operational": 4, 00:09:37.988 "base_bdevs_list": [ 00:09:37.988 { 00:09:37.988 "name": "BaseBdev1", 00:09:37.988 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:37.988 "is_configured": false, 00:09:37.988 "data_offset": 0, 00:09:37.988 "data_size": 0 00:09:37.988 }, 00:09:37.988 { 00:09:37.988 "name": "BaseBdev2", 00:09:37.988 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:37.988 "is_configured": false, 00:09:37.988 "data_offset": 0, 00:09:37.988 "data_size": 0 00:09:37.988 }, 00:09:37.988 { 00:09:37.988 "name": "BaseBdev3", 00:09:37.988 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:37.988 "is_configured": false, 00:09:37.988 "data_offset": 0, 00:09:37.988 "data_size": 0 00:09:37.988 }, 00:09:37.988 { 00:09:37.988 "name": "BaseBdev4", 00:09:37.988 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:37.988 "is_configured": false, 00:09:37.988 "data_offset": 0, 00:09:37.988 "data_size": 0 00:09:37.988 } 00:09:37.988 ] 00:09:37.988 }' 00:09:37.989 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:37.989 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.248 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:38.248 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.248 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.248 [2024-12-05 18:59:55.678450] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:38.248 [2024-12-05 18:59:55.678541] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.249 [2024-12-05 18:59:55.686461] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:38.249 [2024-12-05 18:59:55.686541] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:38.249 [2024-12-05 18:59:55.686568] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:38.249 [2024-12-05 18:59:55.686590] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:38.249 [2024-12-05 18:59:55.686607] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:38.249 [2024-12-05 18:59:55.686627] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:38.249 [2024-12-05 18:59:55.686644] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:38.249 [2024-12-05 18:59:55.686672] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.249 [2024-12-05 18:59:55.707333] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:38.249 BaseBdev1 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.249 [ 00:09:38.249 { 00:09:38.249 "name": "BaseBdev1", 00:09:38.249 "aliases": [ 00:09:38.249 "79a51fb9-d4f1-4e23-afd5-2a5978825b3f" 00:09:38.249 ], 00:09:38.249 "product_name": "Malloc disk", 00:09:38.249 "block_size": 512, 00:09:38.249 "num_blocks": 65536, 00:09:38.249 "uuid": "79a51fb9-d4f1-4e23-afd5-2a5978825b3f", 00:09:38.249 "assigned_rate_limits": { 00:09:38.249 "rw_ios_per_sec": 0, 00:09:38.249 "rw_mbytes_per_sec": 0, 00:09:38.249 "r_mbytes_per_sec": 0, 00:09:38.249 "w_mbytes_per_sec": 0 00:09:38.249 }, 00:09:38.249 "claimed": true, 00:09:38.249 "claim_type": "exclusive_write", 00:09:38.249 "zoned": false, 00:09:38.249 "supported_io_types": { 00:09:38.249 "read": true, 00:09:38.249 "write": true, 00:09:38.249 "unmap": true, 00:09:38.249 "flush": true, 00:09:38.249 "reset": true, 00:09:38.249 "nvme_admin": false, 00:09:38.249 "nvme_io": false, 00:09:38.249 "nvme_io_md": false, 00:09:38.249 "write_zeroes": true, 00:09:38.249 "zcopy": true, 00:09:38.249 "get_zone_info": false, 00:09:38.249 "zone_management": false, 00:09:38.249 "zone_append": false, 00:09:38.249 "compare": false, 00:09:38.249 "compare_and_write": false, 00:09:38.249 "abort": true, 00:09:38.249 "seek_hole": false, 00:09:38.249 "seek_data": false, 00:09:38.249 "copy": true, 00:09:38.249 "nvme_iov_md": false 00:09:38.249 }, 00:09:38.249 "memory_domains": [ 00:09:38.249 { 00:09:38.249 "dma_device_id": "system", 00:09:38.249 "dma_device_type": 1 00:09:38.249 }, 00:09:38.249 { 00:09:38.249 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:38.249 "dma_device_type": 2 00:09:38.249 } 00:09:38.249 ], 00:09:38.249 "driver_specific": {} 00:09:38.249 } 00:09:38.249 ] 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:38.249 "name": "Existed_Raid", 00:09:38.249 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:38.249 "strip_size_kb": 64, 00:09:38.249 "state": "configuring", 00:09:38.249 "raid_level": "raid0", 00:09:38.249 "superblock": false, 00:09:38.249 "num_base_bdevs": 4, 00:09:38.249 "num_base_bdevs_discovered": 1, 00:09:38.249 "num_base_bdevs_operational": 4, 00:09:38.249 "base_bdevs_list": [ 00:09:38.249 { 00:09:38.249 "name": "BaseBdev1", 00:09:38.249 "uuid": "79a51fb9-d4f1-4e23-afd5-2a5978825b3f", 00:09:38.249 "is_configured": true, 00:09:38.249 "data_offset": 0, 00:09:38.249 "data_size": 65536 00:09:38.249 }, 00:09:38.249 { 00:09:38.249 "name": "BaseBdev2", 00:09:38.249 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:38.249 "is_configured": false, 00:09:38.249 "data_offset": 0, 00:09:38.249 "data_size": 0 00:09:38.249 }, 00:09:38.249 { 00:09:38.249 "name": "BaseBdev3", 00:09:38.249 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:38.249 "is_configured": false, 00:09:38.249 "data_offset": 0, 00:09:38.249 "data_size": 0 00:09:38.249 }, 00:09:38.249 { 00:09:38.249 "name": "BaseBdev4", 00:09:38.249 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:38.249 "is_configured": false, 00:09:38.249 "data_offset": 0, 00:09:38.249 "data_size": 0 00:09:38.249 } 00:09:38.249 ] 00:09:38.249 }' 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:38.249 18:59:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.822 [2024-12-05 18:59:56.198532] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:38.822 [2024-12-05 18:59:56.198637] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.822 [2024-12-05 18:59:56.206557] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:38.822 [2024-12-05 18:59:56.208409] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:38.822 [2024-12-05 18:59:56.208451] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:38.822 [2024-12-05 18:59:56.208461] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:38.822 [2024-12-05 18:59:56.208485] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:38.822 [2024-12-05 18:59:56.208492] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:38.822 [2024-12-05 18:59:56.208500] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:38.822 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:38.822 "name": "Existed_Raid", 00:09:38.822 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:38.822 "strip_size_kb": 64, 00:09:38.822 "state": "configuring", 00:09:38.822 "raid_level": "raid0", 00:09:38.822 "superblock": false, 00:09:38.822 "num_base_bdevs": 4, 00:09:38.822 "num_base_bdevs_discovered": 1, 00:09:38.822 "num_base_bdevs_operational": 4, 00:09:38.822 "base_bdevs_list": [ 00:09:38.822 { 00:09:38.822 "name": "BaseBdev1", 00:09:38.822 "uuid": "79a51fb9-d4f1-4e23-afd5-2a5978825b3f", 00:09:38.822 "is_configured": true, 00:09:38.822 "data_offset": 0, 00:09:38.822 "data_size": 65536 00:09:38.822 }, 00:09:38.822 { 00:09:38.822 "name": "BaseBdev2", 00:09:38.822 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:38.822 "is_configured": false, 00:09:38.822 "data_offset": 0, 00:09:38.822 "data_size": 0 00:09:38.822 }, 00:09:38.822 { 00:09:38.822 "name": "BaseBdev3", 00:09:38.822 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:38.822 "is_configured": false, 00:09:38.823 "data_offset": 0, 00:09:38.823 "data_size": 0 00:09:38.823 }, 00:09:38.823 { 00:09:38.823 "name": "BaseBdev4", 00:09:38.823 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:38.823 "is_configured": false, 00:09:38.823 "data_offset": 0, 00:09:38.823 "data_size": 0 00:09:38.823 } 00:09:38.823 ] 00:09:38.823 }' 00:09:38.823 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:38.823 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.085 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:39.085 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.085 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.346 [2024-12-05 18:59:56.644948] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:39.346 BaseBdev2 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.346 [ 00:09:39.346 { 00:09:39.346 "name": "BaseBdev2", 00:09:39.346 "aliases": [ 00:09:39.346 "1adf4f95-3ce8-4023-9618-17434ee61e14" 00:09:39.346 ], 00:09:39.346 "product_name": "Malloc disk", 00:09:39.346 "block_size": 512, 00:09:39.346 "num_blocks": 65536, 00:09:39.346 "uuid": "1adf4f95-3ce8-4023-9618-17434ee61e14", 00:09:39.346 "assigned_rate_limits": { 00:09:39.346 "rw_ios_per_sec": 0, 00:09:39.346 "rw_mbytes_per_sec": 0, 00:09:39.346 "r_mbytes_per_sec": 0, 00:09:39.346 "w_mbytes_per_sec": 0 00:09:39.346 }, 00:09:39.346 "claimed": true, 00:09:39.346 "claim_type": "exclusive_write", 00:09:39.346 "zoned": false, 00:09:39.346 "supported_io_types": { 00:09:39.346 "read": true, 00:09:39.346 "write": true, 00:09:39.346 "unmap": true, 00:09:39.346 "flush": true, 00:09:39.346 "reset": true, 00:09:39.346 "nvme_admin": false, 00:09:39.346 "nvme_io": false, 00:09:39.346 "nvme_io_md": false, 00:09:39.346 "write_zeroes": true, 00:09:39.346 "zcopy": true, 00:09:39.346 "get_zone_info": false, 00:09:39.346 "zone_management": false, 00:09:39.346 "zone_append": false, 00:09:39.346 "compare": false, 00:09:39.346 "compare_and_write": false, 00:09:39.346 "abort": true, 00:09:39.346 "seek_hole": false, 00:09:39.346 "seek_data": false, 00:09:39.346 "copy": true, 00:09:39.346 "nvme_iov_md": false 00:09:39.346 }, 00:09:39.346 "memory_domains": [ 00:09:39.346 { 00:09:39.346 "dma_device_id": "system", 00:09:39.346 "dma_device_type": 1 00:09:39.346 }, 00:09:39.346 { 00:09:39.346 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.346 "dma_device_type": 2 00:09:39.346 } 00:09:39.346 ], 00:09:39.346 "driver_specific": {} 00:09:39.346 } 00:09:39.346 ] 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:39.346 "name": "Existed_Raid", 00:09:39.346 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:39.346 "strip_size_kb": 64, 00:09:39.346 "state": "configuring", 00:09:39.346 "raid_level": "raid0", 00:09:39.346 "superblock": false, 00:09:39.346 "num_base_bdevs": 4, 00:09:39.346 "num_base_bdevs_discovered": 2, 00:09:39.346 "num_base_bdevs_operational": 4, 00:09:39.346 "base_bdevs_list": [ 00:09:39.346 { 00:09:39.346 "name": "BaseBdev1", 00:09:39.346 "uuid": "79a51fb9-d4f1-4e23-afd5-2a5978825b3f", 00:09:39.346 "is_configured": true, 00:09:39.346 "data_offset": 0, 00:09:39.346 "data_size": 65536 00:09:39.346 }, 00:09:39.346 { 00:09:39.346 "name": "BaseBdev2", 00:09:39.346 "uuid": "1adf4f95-3ce8-4023-9618-17434ee61e14", 00:09:39.346 "is_configured": true, 00:09:39.346 "data_offset": 0, 00:09:39.346 "data_size": 65536 00:09:39.346 }, 00:09:39.346 { 00:09:39.346 "name": "BaseBdev3", 00:09:39.346 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:39.346 "is_configured": false, 00:09:39.346 "data_offset": 0, 00:09:39.346 "data_size": 0 00:09:39.346 }, 00:09:39.346 { 00:09:39.346 "name": "BaseBdev4", 00:09:39.346 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:39.346 "is_configured": false, 00:09:39.346 "data_offset": 0, 00:09:39.346 "data_size": 0 00:09:39.346 } 00:09:39.346 ] 00:09:39.346 }' 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:39.346 18:59:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.608 [2024-12-05 18:59:57.097060] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:39.608 BaseBdev3 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.608 [ 00:09:39.608 { 00:09:39.608 "name": "BaseBdev3", 00:09:39.608 "aliases": [ 00:09:39.608 "171e8c97-22d3-43d5-8027-e7190b9ea49d" 00:09:39.608 ], 00:09:39.608 "product_name": "Malloc disk", 00:09:39.608 "block_size": 512, 00:09:39.608 "num_blocks": 65536, 00:09:39.608 "uuid": "171e8c97-22d3-43d5-8027-e7190b9ea49d", 00:09:39.608 "assigned_rate_limits": { 00:09:39.608 "rw_ios_per_sec": 0, 00:09:39.608 "rw_mbytes_per_sec": 0, 00:09:39.608 "r_mbytes_per_sec": 0, 00:09:39.608 "w_mbytes_per_sec": 0 00:09:39.608 }, 00:09:39.608 "claimed": true, 00:09:39.608 "claim_type": "exclusive_write", 00:09:39.608 "zoned": false, 00:09:39.608 "supported_io_types": { 00:09:39.608 "read": true, 00:09:39.608 "write": true, 00:09:39.608 "unmap": true, 00:09:39.608 "flush": true, 00:09:39.608 "reset": true, 00:09:39.608 "nvme_admin": false, 00:09:39.608 "nvme_io": false, 00:09:39.608 "nvme_io_md": false, 00:09:39.608 "write_zeroes": true, 00:09:39.608 "zcopy": true, 00:09:39.608 "get_zone_info": false, 00:09:39.608 "zone_management": false, 00:09:39.608 "zone_append": false, 00:09:39.608 "compare": false, 00:09:39.608 "compare_and_write": false, 00:09:39.608 "abort": true, 00:09:39.608 "seek_hole": false, 00:09:39.608 "seek_data": false, 00:09:39.608 "copy": true, 00:09:39.608 "nvme_iov_md": false 00:09:39.608 }, 00:09:39.608 "memory_domains": [ 00:09:39.608 { 00:09:39.608 "dma_device_id": "system", 00:09:39.608 "dma_device_type": 1 00:09:39.608 }, 00:09:39.608 { 00:09:39.608 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.608 "dma_device_type": 2 00:09:39.608 } 00:09:39.608 ], 00:09:39.608 "driver_specific": {} 00:09:39.608 } 00:09:39.608 ] 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.608 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:39.868 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:39.868 "name": "Existed_Raid", 00:09:39.868 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:39.868 "strip_size_kb": 64, 00:09:39.868 "state": "configuring", 00:09:39.868 "raid_level": "raid0", 00:09:39.868 "superblock": false, 00:09:39.868 "num_base_bdevs": 4, 00:09:39.868 "num_base_bdevs_discovered": 3, 00:09:39.868 "num_base_bdevs_operational": 4, 00:09:39.868 "base_bdevs_list": [ 00:09:39.868 { 00:09:39.868 "name": "BaseBdev1", 00:09:39.868 "uuid": "79a51fb9-d4f1-4e23-afd5-2a5978825b3f", 00:09:39.868 "is_configured": true, 00:09:39.868 "data_offset": 0, 00:09:39.868 "data_size": 65536 00:09:39.868 }, 00:09:39.868 { 00:09:39.868 "name": "BaseBdev2", 00:09:39.868 "uuid": "1adf4f95-3ce8-4023-9618-17434ee61e14", 00:09:39.868 "is_configured": true, 00:09:39.868 "data_offset": 0, 00:09:39.868 "data_size": 65536 00:09:39.868 }, 00:09:39.868 { 00:09:39.868 "name": "BaseBdev3", 00:09:39.868 "uuid": "171e8c97-22d3-43d5-8027-e7190b9ea49d", 00:09:39.868 "is_configured": true, 00:09:39.868 "data_offset": 0, 00:09:39.868 "data_size": 65536 00:09:39.868 }, 00:09:39.868 { 00:09:39.868 "name": "BaseBdev4", 00:09:39.868 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:39.868 "is_configured": false, 00:09:39.868 "data_offset": 0, 00:09:39.868 "data_size": 0 00:09:39.868 } 00:09:39.868 ] 00:09:39.868 }' 00:09:39.868 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:39.868 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.127 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.128 [2024-12-05 18:59:57.591284] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:40.128 [2024-12-05 18:59:57.591329] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:40.128 [2024-12-05 18:59:57.591338] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:40.128 [2024-12-05 18:59:57.591647] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:40.128 [2024-12-05 18:59:57.591816] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:40.128 [2024-12-05 18:59:57.591831] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:40.128 [2024-12-05 18:59:57.592035] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:40.128 BaseBdev4 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.128 [ 00:09:40.128 { 00:09:40.128 "name": "BaseBdev4", 00:09:40.128 "aliases": [ 00:09:40.128 "37998a31-0bf3-4684-a1f8-f0d24db1025e" 00:09:40.128 ], 00:09:40.128 "product_name": "Malloc disk", 00:09:40.128 "block_size": 512, 00:09:40.128 "num_blocks": 65536, 00:09:40.128 "uuid": "37998a31-0bf3-4684-a1f8-f0d24db1025e", 00:09:40.128 "assigned_rate_limits": { 00:09:40.128 "rw_ios_per_sec": 0, 00:09:40.128 "rw_mbytes_per_sec": 0, 00:09:40.128 "r_mbytes_per_sec": 0, 00:09:40.128 "w_mbytes_per_sec": 0 00:09:40.128 }, 00:09:40.128 "claimed": true, 00:09:40.128 "claim_type": "exclusive_write", 00:09:40.128 "zoned": false, 00:09:40.128 "supported_io_types": { 00:09:40.128 "read": true, 00:09:40.128 "write": true, 00:09:40.128 "unmap": true, 00:09:40.128 "flush": true, 00:09:40.128 "reset": true, 00:09:40.128 "nvme_admin": false, 00:09:40.128 "nvme_io": false, 00:09:40.128 "nvme_io_md": false, 00:09:40.128 "write_zeroes": true, 00:09:40.128 "zcopy": true, 00:09:40.128 "get_zone_info": false, 00:09:40.128 "zone_management": false, 00:09:40.128 "zone_append": false, 00:09:40.128 "compare": false, 00:09:40.128 "compare_and_write": false, 00:09:40.128 "abort": true, 00:09:40.128 "seek_hole": false, 00:09:40.128 "seek_data": false, 00:09:40.128 "copy": true, 00:09:40.128 "nvme_iov_md": false 00:09:40.128 }, 00:09:40.128 "memory_domains": [ 00:09:40.128 { 00:09:40.128 "dma_device_id": "system", 00:09:40.128 "dma_device_type": 1 00:09:40.128 }, 00:09:40.128 { 00:09:40.128 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:40.128 "dma_device_type": 2 00:09:40.128 } 00:09:40.128 ], 00:09:40.128 "driver_specific": {} 00:09:40.128 } 00:09:40.128 ] 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:40.128 "name": "Existed_Raid", 00:09:40.128 "uuid": "fa2152fb-78f5-49d8-89d0-9d4c07df9d90", 00:09:40.128 "strip_size_kb": 64, 00:09:40.128 "state": "online", 00:09:40.128 "raid_level": "raid0", 00:09:40.128 "superblock": false, 00:09:40.128 "num_base_bdevs": 4, 00:09:40.128 "num_base_bdevs_discovered": 4, 00:09:40.128 "num_base_bdevs_operational": 4, 00:09:40.128 "base_bdevs_list": [ 00:09:40.128 { 00:09:40.128 "name": "BaseBdev1", 00:09:40.128 "uuid": "79a51fb9-d4f1-4e23-afd5-2a5978825b3f", 00:09:40.128 "is_configured": true, 00:09:40.128 "data_offset": 0, 00:09:40.128 "data_size": 65536 00:09:40.128 }, 00:09:40.128 { 00:09:40.128 "name": "BaseBdev2", 00:09:40.128 "uuid": "1adf4f95-3ce8-4023-9618-17434ee61e14", 00:09:40.128 "is_configured": true, 00:09:40.128 "data_offset": 0, 00:09:40.128 "data_size": 65536 00:09:40.128 }, 00:09:40.128 { 00:09:40.128 "name": "BaseBdev3", 00:09:40.128 "uuid": "171e8c97-22d3-43d5-8027-e7190b9ea49d", 00:09:40.128 "is_configured": true, 00:09:40.128 "data_offset": 0, 00:09:40.128 "data_size": 65536 00:09:40.128 }, 00:09:40.128 { 00:09:40.128 "name": "BaseBdev4", 00:09:40.128 "uuid": "37998a31-0bf3-4684-a1f8-f0d24db1025e", 00:09:40.128 "is_configured": true, 00:09:40.128 "data_offset": 0, 00:09:40.128 "data_size": 65536 00:09:40.128 } 00:09:40.128 ] 00:09:40.128 }' 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:40.128 18:59:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.698 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:40.698 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:40.698 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:40.698 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:40.698 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:40.698 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:40.698 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:40.698 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.698 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.698 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:40.698 [2024-12-05 18:59:58.086824] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:40.698 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.698 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:40.698 "name": "Existed_Raid", 00:09:40.698 "aliases": [ 00:09:40.698 "fa2152fb-78f5-49d8-89d0-9d4c07df9d90" 00:09:40.698 ], 00:09:40.698 "product_name": "Raid Volume", 00:09:40.698 "block_size": 512, 00:09:40.698 "num_blocks": 262144, 00:09:40.698 "uuid": "fa2152fb-78f5-49d8-89d0-9d4c07df9d90", 00:09:40.698 "assigned_rate_limits": { 00:09:40.698 "rw_ios_per_sec": 0, 00:09:40.698 "rw_mbytes_per_sec": 0, 00:09:40.698 "r_mbytes_per_sec": 0, 00:09:40.698 "w_mbytes_per_sec": 0 00:09:40.698 }, 00:09:40.698 "claimed": false, 00:09:40.698 "zoned": false, 00:09:40.698 "supported_io_types": { 00:09:40.698 "read": true, 00:09:40.698 "write": true, 00:09:40.698 "unmap": true, 00:09:40.698 "flush": true, 00:09:40.698 "reset": true, 00:09:40.698 "nvme_admin": false, 00:09:40.698 "nvme_io": false, 00:09:40.698 "nvme_io_md": false, 00:09:40.698 "write_zeroes": true, 00:09:40.698 "zcopy": false, 00:09:40.698 "get_zone_info": false, 00:09:40.698 "zone_management": false, 00:09:40.698 "zone_append": false, 00:09:40.698 "compare": false, 00:09:40.698 "compare_and_write": false, 00:09:40.698 "abort": false, 00:09:40.698 "seek_hole": false, 00:09:40.698 "seek_data": false, 00:09:40.698 "copy": false, 00:09:40.698 "nvme_iov_md": false 00:09:40.698 }, 00:09:40.698 "memory_domains": [ 00:09:40.698 { 00:09:40.698 "dma_device_id": "system", 00:09:40.698 "dma_device_type": 1 00:09:40.698 }, 00:09:40.698 { 00:09:40.698 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:40.698 "dma_device_type": 2 00:09:40.698 }, 00:09:40.698 { 00:09:40.698 "dma_device_id": "system", 00:09:40.698 "dma_device_type": 1 00:09:40.698 }, 00:09:40.698 { 00:09:40.698 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:40.698 "dma_device_type": 2 00:09:40.698 }, 00:09:40.698 { 00:09:40.698 "dma_device_id": "system", 00:09:40.698 "dma_device_type": 1 00:09:40.698 }, 00:09:40.698 { 00:09:40.698 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:40.698 "dma_device_type": 2 00:09:40.698 }, 00:09:40.698 { 00:09:40.698 "dma_device_id": "system", 00:09:40.698 "dma_device_type": 1 00:09:40.698 }, 00:09:40.698 { 00:09:40.698 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:40.698 "dma_device_type": 2 00:09:40.698 } 00:09:40.698 ], 00:09:40.698 "driver_specific": { 00:09:40.698 "raid": { 00:09:40.698 "uuid": "fa2152fb-78f5-49d8-89d0-9d4c07df9d90", 00:09:40.698 "strip_size_kb": 64, 00:09:40.698 "state": "online", 00:09:40.698 "raid_level": "raid0", 00:09:40.698 "superblock": false, 00:09:40.698 "num_base_bdevs": 4, 00:09:40.698 "num_base_bdevs_discovered": 4, 00:09:40.698 "num_base_bdevs_operational": 4, 00:09:40.698 "base_bdevs_list": [ 00:09:40.698 { 00:09:40.698 "name": "BaseBdev1", 00:09:40.698 "uuid": "79a51fb9-d4f1-4e23-afd5-2a5978825b3f", 00:09:40.698 "is_configured": true, 00:09:40.698 "data_offset": 0, 00:09:40.698 "data_size": 65536 00:09:40.698 }, 00:09:40.698 { 00:09:40.698 "name": "BaseBdev2", 00:09:40.698 "uuid": "1adf4f95-3ce8-4023-9618-17434ee61e14", 00:09:40.698 "is_configured": true, 00:09:40.698 "data_offset": 0, 00:09:40.698 "data_size": 65536 00:09:40.698 }, 00:09:40.699 { 00:09:40.699 "name": "BaseBdev3", 00:09:40.699 "uuid": "171e8c97-22d3-43d5-8027-e7190b9ea49d", 00:09:40.699 "is_configured": true, 00:09:40.699 "data_offset": 0, 00:09:40.699 "data_size": 65536 00:09:40.699 }, 00:09:40.699 { 00:09:40.699 "name": "BaseBdev4", 00:09:40.699 "uuid": "37998a31-0bf3-4684-a1f8-f0d24db1025e", 00:09:40.699 "is_configured": true, 00:09:40.699 "data_offset": 0, 00:09:40.699 "data_size": 65536 00:09:40.699 } 00:09:40.699 ] 00:09:40.699 } 00:09:40.699 } 00:09:40.699 }' 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:40.699 BaseBdev2 00:09:40.699 BaseBdev3 00:09:40.699 BaseBdev4' 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.699 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.959 [2024-12-05 18:59:58.394030] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:40.959 [2024-12-05 18:59:58.394061] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:40.959 [2024-12-05 18:59:58.394113] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:40.959 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.960 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:40.960 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:40.960 "name": "Existed_Raid", 00:09:40.960 "uuid": "fa2152fb-78f5-49d8-89d0-9d4c07df9d90", 00:09:40.960 "strip_size_kb": 64, 00:09:40.960 "state": "offline", 00:09:40.960 "raid_level": "raid0", 00:09:40.960 "superblock": false, 00:09:40.960 "num_base_bdevs": 4, 00:09:40.960 "num_base_bdevs_discovered": 3, 00:09:40.960 "num_base_bdevs_operational": 3, 00:09:40.960 "base_bdevs_list": [ 00:09:40.960 { 00:09:40.960 "name": null, 00:09:40.960 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:40.960 "is_configured": false, 00:09:40.960 "data_offset": 0, 00:09:40.960 "data_size": 65536 00:09:40.960 }, 00:09:40.960 { 00:09:40.960 "name": "BaseBdev2", 00:09:40.960 "uuid": "1adf4f95-3ce8-4023-9618-17434ee61e14", 00:09:40.960 "is_configured": true, 00:09:40.960 "data_offset": 0, 00:09:40.960 "data_size": 65536 00:09:40.960 }, 00:09:40.960 { 00:09:40.960 "name": "BaseBdev3", 00:09:40.960 "uuid": "171e8c97-22d3-43d5-8027-e7190b9ea49d", 00:09:40.960 "is_configured": true, 00:09:40.960 "data_offset": 0, 00:09:40.960 "data_size": 65536 00:09:40.960 }, 00:09:40.960 { 00:09:40.960 "name": "BaseBdev4", 00:09:40.960 "uuid": "37998a31-0bf3-4684-a1f8-f0d24db1025e", 00:09:40.960 "is_configured": true, 00:09:40.960 "data_offset": 0, 00:09:40.960 "data_size": 65536 00:09:40.960 } 00:09:40.960 ] 00:09:40.960 }' 00:09:40.960 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:40.960 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.220 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:41.220 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.482 [2024-12-05 18:59:58.836547] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.482 [2024-12-05 18:59:58.903695] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.482 [2024-12-05 18:59:58.946878] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:41.482 [2024-12-05 18:59:58.946983] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.482 18:59:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.482 BaseBdev2 00:09:41.482 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.482 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:41.482 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:41.482 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:41.482 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:41.482 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:41.482 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:41.482 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:41.482 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.482 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.482 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.482 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:41.482 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.482 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.482 [ 00:09:41.482 { 00:09:41.482 "name": "BaseBdev2", 00:09:41.482 "aliases": [ 00:09:41.482 "dddb5b14-3acd-42bf-8add-d2bf24256080" 00:09:41.482 ], 00:09:41.482 "product_name": "Malloc disk", 00:09:41.482 "block_size": 512, 00:09:41.482 "num_blocks": 65536, 00:09:41.482 "uuid": "dddb5b14-3acd-42bf-8add-d2bf24256080", 00:09:41.482 "assigned_rate_limits": { 00:09:41.482 "rw_ios_per_sec": 0, 00:09:41.482 "rw_mbytes_per_sec": 0, 00:09:41.482 "r_mbytes_per_sec": 0, 00:09:41.482 "w_mbytes_per_sec": 0 00:09:41.482 }, 00:09:41.482 "claimed": false, 00:09:41.482 "zoned": false, 00:09:41.482 "supported_io_types": { 00:09:41.482 "read": true, 00:09:41.482 "write": true, 00:09:41.482 "unmap": true, 00:09:41.482 "flush": true, 00:09:41.482 "reset": true, 00:09:41.482 "nvme_admin": false, 00:09:41.482 "nvme_io": false, 00:09:41.482 "nvme_io_md": false, 00:09:41.482 "write_zeroes": true, 00:09:41.482 "zcopy": true, 00:09:41.482 "get_zone_info": false, 00:09:41.482 "zone_management": false, 00:09:41.482 "zone_append": false, 00:09:41.482 "compare": false, 00:09:41.482 "compare_and_write": false, 00:09:41.482 "abort": true, 00:09:41.482 "seek_hole": false, 00:09:41.482 "seek_data": false, 00:09:41.482 "copy": true, 00:09:41.482 "nvme_iov_md": false 00:09:41.482 }, 00:09:41.482 "memory_domains": [ 00:09:41.482 { 00:09:41.743 "dma_device_id": "system", 00:09:41.743 "dma_device_type": 1 00:09:41.743 }, 00:09:41.743 { 00:09:41.743 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:41.743 "dma_device_type": 2 00:09:41.743 } 00:09:41.743 ], 00:09:41.743 "driver_specific": {} 00:09:41.743 } 00:09:41.743 ] 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.743 BaseBdev3 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.743 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.743 [ 00:09:41.743 { 00:09:41.743 "name": "BaseBdev3", 00:09:41.743 "aliases": [ 00:09:41.743 "cf50d885-709a-4514-8f21-48a4b912c3b8" 00:09:41.743 ], 00:09:41.743 "product_name": "Malloc disk", 00:09:41.744 "block_size": 512, 00:09:41.744 "num_blocks": 65536, 00:09:41.744 "uuid": "cf50d885-709a-4514-8f21-48a4b912c3b8", 00:09:41.744 "assigned_rate_limits": { 00:09:41.744 "rw_ios_per_sec": 0, 00:09:41.744 "rw_mbytes_per_sec": 0, 00:09:41.744 "r_mbytes_per_sec": 0, 00:09:41.744 "w_mbytes_per_sec": 0 00:09:41.744 }, 00:09:41.744 "claimed": false, 00:09:41.744 "zoned": false, 00:09:41.744 "supported_io_types": { 00:09:41.744 "read": true, 00:09:41.744 "write": true, 00:09:41.744 "unmap": true, 00:09:41.744 "flush": true, 00:09:41.744 "reset": true, 00:09:41.744 "nvme_admin": false, 00:09:41.744 "nvme_io": false, 00:09:41.744 "nvme_io_md": false, 00:09:41.744 "write_zeroes": true, 00:09:41.744 "zcopy": true, 00:09:41.744 "get_zone_info": false, 00:09:41.744 "zone_management": false, 00:09:41.744 "zone_append": false, 00:09:41.744 "compare": false, 00:09:41.744 "compare_and_write": false, 00:09:41.744 "abort": true, 00:09:41.744 "seek_hole": false, 00:09:41.744 "seek_data": false, 00:09:41.744 "copy": true, 00:09:41.744 "nvme_iov_md": false 00:09:41.744 }, 00:09:41.744 "memory_domains": [ 00:09:41.744 { 00:09:41.744 "dma_device_id": "system", 00:09:41.744 "dma_device_type": 1 00:09:41.744 }, 00:09:41.744 { 00:09:41.744 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:41.744 "dma_device_type": 2 00:09:41.744 } 00:09:41.744 ], 00:09:41.744 "driver_specific": {} 00:09:41.744 } 00:09:41.744 ] 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.744 BaseBdev4 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.744 [ 00:09:41.744 { 00:09:41.744 "name": "BaseBdev4", 00:09:41.744 "aliases": [ 00:09:41.744 "f156c39e-a303-4a32-8eb8-5af3f3c8cbf3" 00:09:41.744 ], 00:09:41.744 "product_name": "Malloc disk", 00:09:41.744 "block_size": 512, 00:09:41.744 "num_blocks": 65536, 00:09:41.744 "uuid": "f156c39e-a303-4a32-8eb8-5af3f3c8cbf3", 00:09:41.744 "assigned_rate_limits": { 00:09:41.744 "rw_ios_per_sec": 0, 00:09:41.744 "rw_mbytes_per_sec": 0, 00:09:41.744 "r_mbytes_per_sec": 0, 00:09:41.744 "w_mbytes_per_sec": 0 00:09:41.744 }, 00:09:41.744 "claimed": false, 00:09:41.744 "zoned": false, 00:09:41.744 "supported_io_types": { 00:09:41.744 "read": true, 00:09:41.744 "write": true, 00:09:41.744 "unmap": true, 00:09:41.744 "flush": true, 00:09:41.744 "reset": true, 00:09:41.744 "nvme_admin": false, 00:09:41.744 "nvme_io": false, 00:09:41.744 "nvme_io_md": false, 00:09:41.744 "write_zeroes": true, 00:09:41.744 "zcopy": true, 00:09:41.744 "get_zone_info": false, 00:09:41.744 "zone_management": false, 00:09:41.744 "zone_append": false, 00:09:41.744 "compare": false, 00:09:41.744 "compare_and_write": false, 00:09:41.744 "abort": true, 00:09:41.744 "seek_hole": false, 00:09:41.744 "seek_data": false, 00:09:41.744 "copy": true, 00:09:41.744 "nvme_iov_md": false 00:09:41.744 }, 00:09:41.744 "memory_domains": [ 00:09:41.744 { 00:09:41.744 "dma_device_id": "system", 00:09:41.744 "dma_device_type": 1 00:09:41.744 }, 00:09:41.744 { 00:09:41.744 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:41.744 "dma_device_type": 2 00:09:41.744 } 00:09:41.744 ], 00:09:41.744 "driver_specific": {} 00:09:41.744 } 00:09:41.744 ] 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.744 [2024-12-05 18:59:59.154624] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:41.744 [2024-12-05 18:59:59.154725] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:41.744 [2024-12-05 18:59:59.154806] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:41.744 [2024-12-05 18:59:59.156611] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:41.744 [2024-12-05 18:59:59.156727] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:41.744 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:41.744 "name": "Existed_Raid", 00:09:41.744 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:41.744 "strip_size_kb": 64, 00:09:41.744 "state": "configuring", 00:09:41.744 "raid_level": "raid0", 00:09:41.744 "superblock": false, 00:09:41.744 "num_base_bdevs": 4, 00:09:41.744 "num_base_bdevs_discovered": 3, 00:09:41.744 "num_base_bdevs_operational": 4, 00:09:41.744 "base_bdevs_list": [ 00:09:41.744 { 00:09:41.744 "name": "BaseBdev1", 00:09:41.744 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:41.744 "is_configured": false, 00:09:41.744 "data_offset": 0, 00:09:41.744 "data_size": 0 00:09:41.744 }, 00:09:41.744 { 00:09:41.744 "name": "BaseBdev2", 00:09:41.744 "uuid": "dddb5b14-3acd-42bf-8add-d2bf24256080", 00:09:41.744 "is_configured": true, 00:09:41.744 "data_offset": 0, 00:09:41.744 "data_size": 65536 00:09:41.744 }, 00:09:41.744 { 00:09:41.744 "name": "BaseBdev3", 00:09:41.744 "uuid": "cf50d885-709a-4514-8f21-48a4b912c3b8", 00:09:41.744 "is_configured": true, 00:09:41.744 "data_offset": 0, 00:09:41.744 "data_size": 65536 00:09:41.744 }, 00:09:41.744 { 00:09:41.744 "name": "BaseBdev4", 00:09:41.744 "uuid": "f156c39e-a303-4a32-8eb8-5af3f3c8cbf3", 00:09:41.744 "is_configured": true, 00:09:41.745 "data_offset": 0, 00:09:41.745 "data_size": 65536 00:09:41.745 } 00:09:41.745 ] 00:09:41.745 }' 00:09:41.745 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:41.745 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.313 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:42.313 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.313 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.313 [2024-12-05 18:59:59.605853] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:42.313 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.313 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:42.313 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:42.313 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:42.313 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:42.313 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:42.313 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:42.313 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:42.313 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:42.313 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:42.314 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:42.314 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:42.314 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.314 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.314 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:42.314 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.314 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:42.314 "name": "Existed_Raid", 00:09:42.314 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:42.314 "strip_size_kb": 64, 00:09:42.314 "state": "configuring", 00:09:42.314 "raid_level": "raid0", 00:09:42.314 "superblock": false, 00:09:42.314 "num_base_bdevs": 4, 00:09:42.314 "num_base_bdevs_discovered": 2, 00:09:42.314 "num_base_bdevs_operational": 4, 00:09:42.314 "base_bdevs_list": [ 00:09:42.314 { 00:09:42.314 "name": "BaseBdev1", 00:09:42.314 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:42.314 "is_configured": false, 00:09:42.314 "data_offset": 0, 00:09:42.314 "data_size": 0 00:09:42.314 }, 00:09:42.314 { 00:09:42.314 "name": null, 00:09:42.314 "uuid": "dddb5b14-3acd-42bf-8add-d2bf24256080", 00:09:42.314 "is_configured": false, 00:09:42.314 "data_offset": 0, 00:09:42.314 "data_size": 65536 00:09:42.314 }, 00:09:42.314 { 00:09:42.314 "name": "BaseBdev3", 00:09:42.314 "uuid": "cf50d885-709a-4514-8f21-48a4b912c3b8", 00:09:42.314 "is_configured": true, 00:09:42.314 "data_offset": 0, 00:09:42.314 "data_size": 65536 00:09:42.314 }, 00:09:42.314 { 00:09:42.314 "name": "BaseBdev4", 00:09:42.314 "uuid": "f156c39e-a303-4a32-8eb8-5af3f3c8cbf3", 00:09:42.314 "is_configured": true, 00:09:42.314 "data_offset": 0, 00:09:42.314 "data_size": 65536 00:09:42.314 } 00:09:42.314 ] 00:09:42.314 }' 00:09:42.314 18:59:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:42.314 18:59:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.573 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.574 [2024-12-05 19:00:00.064086] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:42.574 BaseBdev1 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.574 [ 00:09:42.574 { 00:09:42.574 "name": "BaseBdev1", 00:09:42.574 "aliases": [ 00:09:42.574 "7be174e6-d485-4540-8579-5ae204c00bf5" 00:09:42.574 ], 00:09:42.574 "product_name": "Malloc disk", 00:09:42.574 "block_size": 512, 00:09:42.574 "num_blocks": 65536, 00:09:42.574 "uuid": "7be174e6-d485-4540-8579-5ae204c00bf5", 00:09:42.574 "assigned_rate_limits": { 00:09:42.574 "rw_ios_per_sec": 0, 00:09:42.574 "rw_mbytes_per_sec": 0, 00:09:42.574 "r_mbytes_per_sec": 0, 00:09:42.574 "w_mbytes_per_sec": 0 00:09:42.574 }, 00:09:42.574 "claimed": true, 00:09:42.574 "claim_type": "exclusive_write", 00:09:42.574 "zoned": false, 00:09:42.574 "supported_io_types": { 00:09:42.574 "read": true, 00:09:42.574 "write": true, 00:09:42.574 "unmap": true, 00:09:42.574 "flush": true, 00:09:42.574 "reset": true, 00:09:42.574 "nvme_admin": false, 00:09:42.574 "nvme_io": false, 00:09:42.574 "nvme_io_md": false, 00:09:42.574 "write_zeroes": true, 00:09:42.574 "zcopy": true, 00:09:42.574 "get_zone_info": false, 00:09:42.574 "zone_management": false, 00:09:42.574 "zone_append": false, 00:09:42.574 "compare": false, 00:09:42.574 "compare_and_write": false, 00:09:42.574 "abort": true, 00:09:42.574 "seek_hole": false, 00:09:42.574 "seek_data": false, 00:09:42.574 "copy": true, 00:09:42.574 "nvme_iov_md": false 00:09:42.574 }, 00:09:42.574 "memory_domains": [ 00:09:42.574 { 00:09:42.574 "dma_device_id": "system", 00:09:42.574 "dma_device_type": 1 00:09:42.574 }, 00:09:42.574 { 00:09:42.574 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:42.574 "dma_device_type": 2 00:09:42.574 } 00:09:42.574 ], 00:09:42.574 "driver_specific": {} 00:09:42.574 } 00:09:42.574 ] 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:42.574 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:42.833 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:42.833 "name": "Existed_Raid", 00:09:42.833 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:42.833 "strip_size_kb": 64, 00:09:42.833 "state": "configuring", 00:09:42.833 "raid_level": "raid0", 00:09:42.833 "superblock": false, 00:09:42.833 "num_base_bdevs": 4, 00:09:42.833 "num_base_bdevs_discovered": 3, 00:09:42.833 "num_base_bdevs_operational": 4, 00:09:42.833 "base_bdevs_list": [ 00:09:42.833 { 00:09:42.833 "name": "BaseBdev1", 00:09:42.833 "uuid": "7be174e6-d485-4540-8579-5ae204c00bf5", 00:09:42.833 "is_configured": true, 00:09:42.833 "data_offset": 0, 00:09:42.833 "data_size": 65536 00:09:42.833 }, 00:09:42.833 { 00:09:42.833 "name": null, 00:09:42.833 "uuid": "dddb5b14-3acd-42bf-8add-d2bf24256080", 00:09:42.833 "is_configured": false, 00:09:42.833 "data_offset": 0, 00:09:42.833 "data_size": 65536 00:09:42.833 }, 00:09:42.833 { 00:09:42.833 "name": "BaseBdev3", 00:09:42.833 "uuid": "cf50d885-709a-4514-8f21-48a4b912c3b8", 00:09:42.833 "is_configured": true, 00:09:42.833 "data_offset": 0, 00:09:42.833 "data_size": 65536 00:09:42.833 }, 00:09:42.833 { 00:09:42.833 "name": "BaseBdev4", 00:09:42.833 "uuid": "f156c39e-a303-4a32-8eb8-5af3f3c8cbf3", 00:09:42.834 "is_configured": true, 00:09:42.834 "data_offset": 0, 00:09:42.834 "data_size": 65536 00:09:42.834 } 00:09:42.834 ] 00:09:42.834 }' 00:09:42.834 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:42.834 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.094 [2024-12-05 19:00:00.507426] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:43.094 "name": "Existed_Raid", 00:09:43.094 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:43.094 "strip_size_kb": 64, 00:09:43.094 "state": "configuring", 00:09:43.094 "raid_level": "raid0", 00:09:43.094 "superblock": false, 00:09:43.094 "num_base_bdevs": 4, 00:09:43.094 "num_base_bdevs_discovered": 2, 00:09:43.094 "num_base_bdevs_operational": 4, 00:09:43.094 "base_bdevs_list": [ 00:09:43.094 { 00:09:43.094 "name": "BaseBdev1", 00:09:43.094 "uuid": "7be174e6-d485-4540-8579-5ae204c00bf5", 00:09:43.094 "is_configured": true, 00:09:43.094 "data_offset": 0, 00:09:43.094 "data_size": 65536 00:09:43.094 }, 00:09:43.094 { 00:09:43.094 "name": null, 00:09:43.094 "uuid": "dddb5b14-3acd-42bf-8add-d2bf24256080", 00:09:43.094 "is_configured": false, 00:09:43.094 "data_offset": 0, 00:09:43.094 "data_size": 65536 00:09:43.094 }, 00:09:43.094 { 00:09:43.094 "name": null, 00:09:43.094 "uuid": "cf50d885-709a-4514-8f21-48a4b912c3b8", 00:09:43.094 "is_configured": false, 00:09:43.094 "data_offset": 0, 00:09:43.094 "data_size": 65536 00:09:43.094 }, 00:09:43.094 { 00:09:43.094 "name": "BaseBdev4", 00:09:43.094 "uuid": "f156c39e-a303-4a32-8eb8-5af3f3c8cbf3", 00:09:43.094 "is_configured": true, 00:09:43.094 "data_offset": 0, 00:09:43.094 "data_size": 65536 00:09:43.094 } 00:09:43.094 ] 00:09:43.094 }' 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:43.094 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.354 [2024-12-05 19:00:00.890768] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:43.354 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.614 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.614 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:43.614 "name": "Existed_Raid", 00:09:43.614 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:43.614 "strip_size_kb": 64, 00:09:43.614 "state": "configuring", 00:09:43.614 "raid_level": "raid0", 00:09:43.614 "superblock": false, 00:09:43.614 "num_base_bdevs": 4, 00:09:43.614 "num_base_bdevs_discovered": 3, 00:09:43.614 "num_base_bdevs_operational": 4, 00:09:43.614 "base_bdevs_list": [ 00:09:43.614 { 00:09:43.614 "name": "BaseBdev1", 00:09:43.614 "uuid": "7be174e6-d485-4540-8579-5ae204c00bf5", 00:09:43.614 "is_configured": true, 00:09:43.614 "data_offset": 0, 00:09:43.614 "data_size": 65536 00:09:43.614 }, 00:09:43.614 { 00:09:43.614 "name": null, 00:09:43.614 "uuid": "dddb5b14-3acd-42bf-8add-d2bf24256080", 00:09:43.614 "is_configured": false, 00:09:43.614 "data_offset": 0, 00:09:43.614 "data_size": 65536 00:09:43.614 }, 00:09:43.614 { 00:09:43.614 "name": "BaseBdev3", 00:09:43.614 "uuid": "cf50d885-709a-4514-8f21-48a4b912c3b8", 00:09:43.614 "is_configured": true, 00:09:43.614 "data_offset": 0, 00:09:43.614 "data_size": 65536 00:09:43.614 }, 00:09:43.614 { 00:09:43.614 "name": "BaseBdev4", 00:09:43.614 "uuid": "f156c39e-a303-4a32-8eb8-5af3f3c8cbf3", 00:09:43.614 "is_configured": true, 00:09:43.614 "data_offset": 0, 00:09:43.614 "data_size": 65536 00:09:43.614 } 00:09:43.614 ] 00:09:43.614 }' 00:09:43.614 19:00:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:43.614 19:00:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.874 [2024-12-05 19:00:01.298094] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:43.874 "name": "Existed_Raid", 00:09:43.874 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:43.874 "strip_size_kb": 64, 00:09:43.874 "state": "configuring", 00:09:43.874 "raid_level": "raid0", 00:09:43.874 "superblock": false, 00:09:43.874 "num_base_bdevs": 4, 00:09:43.874 "num_base_bdevs_discovered": 2, 00:09:43.874 "num_base_bdevs_operational": 4, 00:09:43.874 "base_bdevs_list": [ 00:09:43.874 { 00:09:43.874 "name": null, 00:09:43.874 "uuid": "7be174e6-d485-4540-8579-5ae204c00bf5", 00:09:43.874 "is_configured": false, 00:09:43.874 "data_offset": 0, 00:09:43.874 "data_size": 65536 00:09:43.874 }, 00:09:43.874 { 00:09:43.874 "name": null, 00:09:43.874 "uuid": "dddb5b14-3acd-42bf-8add-d2bf24256080", 00:09:43.874 "is_configured": false, 00:09:43.874 "data_offset": 0, 00:09:43.874 "data_size": 65536 00:09:43.874 }, 00:09:43.874 { 00:09:43.874 "name": "BaseBdev3", 00:09:43.874 "uuid": "cf50d885-709a-4514-8f21-48a4b912c3b8", 00:09:43.874 "is_configured": true, 00:09:43.874 "data_offset": 0, 00:09:43.874 "data_size": 65536 00:09:43.874 }, 00:09:43.874 { 00:09:43.874 "name": "BaseBdev4", 00:09:43.874 "uuid": "f156c39e-a303-4a32-8eb8-5af3f3c8cbf3", 00:09:43.874 "is_configured": true, 00:09:43.874 "data_offset": 0, 00:09:43.874 "data_size": 65536 00:09:43.874 } 00:09:43.874 ] 00:09:43.874 }' 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:43.874 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.134 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.134 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:44.134 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.134 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.134 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.394 [2024-12-05 19:00:01.707887] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.394 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:44.394 "name": "Existed_Raid", 00:09:44.394 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:44.394 "strip_size_kb": 64, 00:09:44.394 "state": "configuring", 00:09:44.394 "raid_level": "raid0", 00:09:44.394 "superblock": false, 00:09:44.394 "num_base_bdevs": 4, 00:09:44.394 "num_base_bdevs_discovered": 3, 00:09:44.394 "num_base_bdevs_operational": 4, 00:09:44.394 "base_bdevs_list": [ 00:09:44.394 { 00:09:44.394 "name": null, 00:09:44.394 "uuid": "7be174e6-d485-4540-8579-5ae204c00bf5", 00:09:44.394 "is_configured": false, 00:09:44.394 "data_offset": 0, 00:09:44.394 "data_size": 65536 00:09:44.394 }, 00:09:44.394 { 00:09:44.394 "name": "BaseBdev2", 00:09:44.394 "uuid": "dddb5b14-3acd-42bf-8add-d2bf24256080", 00:09:44.394 "is_configured": true, 00:09:44.394 "data_offset": 0, 00:09:44.394 "data_size": 65536 00:09:44.394 }, 00:09:44.394 { 00:09:44.394 "name": "BaseBdev3", 00:09:44.394 "uuid": "cf50d885-709a-4514-8f21-48a4b912c3b8", 00:09:44.394 "is_configured": true, 00:09:44.394 "data_offset": 0, 00:09:44.394 "data_size": 65536 00:09:44.394 }, 00:09:44.394 { 00:09:44.394 "name": "BaseBdev4", 00:09:44.395 "uuid": "f156c39e-a303-4a32-8eb8-5af3f3c8cbf3", 00:09:44.395 "is_configured": true, 00:09:44.395 "data_offset": 0, 00:09:44.395 "data_size": 65536 00:09:44.395 } 00:09:44.395 ] 00:09:44.395 }' 00:09:44.395 19:00:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:44.395 19:00:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.654 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.654 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.654 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:44.654 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.654 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.654 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:44.654 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.654 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.654 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.654 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:44.654 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 7be174e6-d485-4540-8579-5ae204c00bf5 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.914 [2024-12-05 19:00:02.241990] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:44.914 [2024-12-05 19:00:02.242112] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:44.914 [2024-12-05 19:00:02.242138] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:44.914 [2024-12-05 19:00:02.242439] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:09:44.914 [2024-12-05 19:00:02.242596] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:44.914 [2024-12-05 19:00:02.242636] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:44.914 [2024-12-05 19:00:02.242871] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:44.914 NewBaseBdev 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.914 [ 00:09:44.914 { 00:09:44.914 "name": "NewBaseBdev", 00:09:44.914 "aliases": [ 00:09:44.914 "7be174e6-d485-4540-8579-5ae204c00bf5" 00:09:44.914 ], 00:09:44.914 "product_name": "Malloc disk", 00:09:44.914 "block_size": 512, 00:09:44.914 "num_blocks": 65536, 00:09:44.914 "uuid": "7be174e6-d485-4540-8579-5ae204c00bf5", 00:09:44.914 "assigned_rate_limits": { 00:09:44.914 "rw_ios_per_sec": 0, 00:09:44.914 "rw_mbytes_per_sec": 0, 00:09:44.914 "r_mbytes_per_sec": 0, 00:09:44.914 "w_mbytes_per_sec": 0 00:09:44.914 }, 00:09:44.914 "claimed": true, 00:09:44.914 "claim_type": "exclusive_write", 00:09:44.914 "zoned": false, 00:09:44.914 "supported_io_types": { 00:09:44.914 "read": true, 00:09:44.914 "write": true, 00:09:44.914 "unmap": true, 00:09:44.914 "flush": true, 00:09:44.914 "reset": true, 00:09:44.914 "nvme_admin": false, 00:09:44.914 "nvme_io": false, 00:09:44.914 "nvme_io_md": false, 00:09:44.914 "write_zeroes": true, 00:09:44.914 "zcopy": true, 00:09:44.914 "get_zone_info": false, 00:09:44.914 "zone_management": false, 00:09:44.914 "zone_append": false, 00:09:44.914 "compare": false, 00:09:44.914 "compare_and_write": false, 00:09:44.914 "abort": true, 00:09:44.914 "seek_hole": false, 00:09:44.914 "seek_data": false, 00:09:44.914 "copy": true, 00:09:44.914 "nvme_iov_md": false 00:09:44.914 }, 00:09:44.914 "memory_domains": [ 00:09:44.914 { 00:09:44.914 "dma_device_id": "system", 00:09:44.914 "dma_device_type": 1 00:09:44.914 }, 00:09:44.914 { 00:09:44.914 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:44.914 "dma_device_type": 2 00:09:44.914 } 00:09:44.914 ], 00:09:44.914 "driver_specific": {} 00:09:44.914 } 00:09:44.914 ] 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:44.914 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:44.914 "name": "Existed_Raid", 00:09:44.914 "uuid": "0a03a8e8-db8f-4b32-8974-0fac5c62627e", 00:09:44.914 "strip_size_kb": 64, 00:09:44.914 "state": "online", 00:09:44.914 "raid_level": "raid0", 00:09:44.914 "superblock": false, 00:09:44.914 "num_base_bdevs": 4, 00:09:44.914 "num_base_bdevs_discovered": 4, 00:09:44.915 "num_base_bdevs_operational": 4, 00:09:44.915 "base_bdevs_list": [ 00:09:44.915 { 00:09:44.915 "name": "NewBaseBdev", 00:09:44.915 "uuid": "7be174e6-d485-4540-8579-5ae204c00bf5", 00:09:44.915 "is_configured": true, 00:09:44.915 "data_offset": 0, 00:09:44.915 "data_size": 65536 00:09:44.915 }, 00:09:44.915 { 00:09:44.915 "name": "BaseBdev2", 00:09:44.915 "uuid": "dddb5b14-3acd-42bf-8add-d2bf24256080", 00:09:44.915 "is_configured": true, 00:09:44.915 "data_offset": 0, 00:09:44.915 "data_size": 65536 00:09:44.915 }, 00:09:44.915 { 00:09:44.915 "name": "BaseBdev3", 00:09:44.915 "uuid": "cf50d885-709a-4514-8f21-48a4b912c3b8", 00:09:44.915 "is_configured": true, 00:09:44.915 "data_offset": 0, 00:09:44.915 "data_size": 65536 00:09:44.915 }, 00:09:44.915 { 00:09:44.915 "name": "BaseBdev4", 00:09:44.915 "uuid": "f156c39e-a303-4a32-8eb8-5af3f3c8cbf3", 00:09:44.915 "is_configured": true, 00:09:44.915 "data_offset": 0, 00:09:44.915 "data_size": 65536 00:09:44.915 } 00:09:44.915 ] 00:09:44.915 }' 00:09:44.915 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:44.915 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.175 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:45.175 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:45.175 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:45.175 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:45.175 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:45.175 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:45.175 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:45.175 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.175 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.175 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:45.175 [2024-12-05 19:00:02.669568] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:45.175 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.175 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:45.175 "name": "Existed_Raid", 00:09:45.175 "aliases": [ 00:09:45.175 "0a03a8e8-db8f-4b32-8974-0fac5c62627e" 00:09:45.175 ], 00:09:45.175 "product_name": "Raid Volume", 00:09:45.175 "block_size": 512, 00:09:45.175 "num_blocks": 262144, 00:09:45.175 "uuid": "0a03a8e8-db8f-4b32-8974-0fac5c62627e", 00:09:45.175 "assigned_rate_limits": { 00:09:45.175 "rw_ios_per_sec": 0, 00:09:45.175 "rw_mbytes_per_sec": 0, 00:09:45.175 "r_mbytes_per_sec": 0, 00:09:45.175 "w_mbytes_per_sec": 0 00:09:45.175 }, 00:09:45.175 "claimed": false, 00:09:45.175 "zoned": false, 00:09:45.175 "supported_io_types": { 00:09:45.175 "read": true, 00:09:45.175 "write": true, 00:09:45.175 "unmap": true, 00:09:45.175 "flush": true, 00:09:45.175 "reset": true, 00:09:45.175 "nvme_admin": false, 00:09:45.175 "nvme_io": false, 00:09:45.175 "nvme_io_md": false, 00:09:45.175 "write_zeroes": true, 00:09:45.175 "zcopy": false, 00:09:45.175 "get_zone_info": false, 00:09:45.175 "zone_management": false, 00:09:45.175 "zone_append": false, 00:09:45.175 "compare": false, 00:09:45.175 "compare_and_write": false, 00:09:45.175 "abort": false, 00:09:45.175 "seek_hole": false, 00:09:45.175 "seek_data": false, 00:09:45.175 "copy": false, 00:09:45.175 "nvme_iov_md": false 00:09:45.175 }, 00:09:45.175 "memory_domains": [ 00:09:45.175 { 00:09:45.175 "dma_device_id": "system", 00:09:45.175 "dma_device_type": 1 00:09:45.175 }, 00:09:45.175 { 00:09:45.175 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:45.175 "dma_device_type": 2 00:09:45.175 }, 00:09:45.175 { 00:09:45.175 "dma_device_id": "system", 00:09:45.175 "dma_device_type": 1 00:09:45.175 }, 00:09:45.175 { 00:09:45.175 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:45.175 "dma_device_type": 2 00:09:45.175 }, 00:09:45.175 { 00:09:45.175 "dma_device_id": "system", 00:09:45.175 "dma_device_type": 1 00:09:45.175 }, 00:09:45.175 { 00:09:45.175 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:45.175 "dma_device_type": 2 00:09:45.175 }, 00:09:45.175 { 00:09:45.175 "dma_device_id": "system", 00:09:45.175 "dma_device_type": 1 00:09:45.175 }, 00:09:45.175 { 00:09:45.175 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:45.175 "dma_device_type": 2 00:09:45.175 } 00:09:45.175 ], 00:09:45.175 "driver_specific": { 00:09:45.175 "raid": { 00:09:45.175 "uuid": "0a03a8e8-db8f-4b32-8974-0fac5c62627e", 00:09:45.175 "strip_size_kb": 64, 00:09:45.175 "state": "online", 00:09:45.175 "raid_level": "raid0", 00:09:45.175 "superblock": false, 00:09:45.175 "num_base_bdevs": 4, 00:09:45.175 "num_base_bdevs_discovered": 4, 00:09:45.175 "num_base_bdevs_operational": 4, 00:09:45.175 "base_bdevs_list": [ 00:09:45.175 { 00:09:45.175 "name": "NewBaseBdev", 00:09:45.175 "uuid": "7be174e6-d485-4540-8579-5ae204c00bf5", 00:09:45.175 "is_configured": true, 00:09:45.175 "data_offset": 0, 00:09:45.175 "data_size": 65536 00:09:45.175 }, 00:09:45.175 { 00:09:45.175 "name": "BaseBdev2", 00:09:45.175 "uuid": "dddb5b14-3acd-42bf-8add-d2bf24256080", 00:09:45.175 "is_configured": true, 00:09:45.175 "data_offset": 0, 00:09:45.175 "data_size": 65536 00:09:45.175 }, 00:09:45.175 { 00:09:45.175 "name": "BaseBdev3", 00:09:45.175 "uuid": "cf50d885-709a-4514-8f21-48a4b912c3b8", 00:09:45.175 "is_configured": true, 00:09:45.175 "data_offset": 0, 00:09:45.175 "data_size": 65536 00:09:45.175 }, 00:09:45.175 { 00:09:45.175 "name": "BaseBdev4", 00:09:45.175 "uuid": "f156c39e-a303-4a32-8eb8-5af3f3c8cbf3", 00:09:45.175 "is_configured": true, 00:09:45.175 "data_offset": 0, 00:09:45.175 "data_size": 65536 00:09:45.176 } 00:09:45.176 ] 00:09:45.176 } 00:09:45.176 } 00:09:45.176 }' 00:09:45.176 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:45.176 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:45.176 BaseBdev2 00:09:45.176 BaseBdev3 00:09:45.176 BaseBdev4' 00:09:45.176 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:45.436 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.437 [2024-12-05 19:00:02.972756] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:45.437 [2024-12-05 19:00:02.972825] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:45.437 [2024-12-05 19:00:02.972908] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:45.437 [2024-12-05 19:00:02.972986] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:45.437 [2024-12-05 19:00:02.973031] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 79679 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 79679 ']' 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 79679 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:45.437 19:00:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 79679 00:09:45.697 19:00:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:45.697 19:00:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:45.697 19:00:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 79679' 00:09:45.697 killing process with pid 79679 00:09:45.697 19:00:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 79679 00:09:45.697 [2024-12-05 19:00:03.019697] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:45.697 19:00:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 79679 00:09:45.697 [2024-12-05 19:00:03.060914] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:45.956 ************************************ 00:09:45.956 END TEST raid_state_function_test 00:09:45.956 ************************************ 00:09:45.956 19:00:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:45.956 00:09:45.956 real 0m8.935s 00:09:45.956 user 0m15.293s 00:09:45.956 sys 0m1.726s 00:09:45.956 19:00:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:45.956 19:00:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.956 19:00:03 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 4 true 00:09:45.956 19:00:03 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:45.956 19:00:03 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:45.956 19:00:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:45.956 ************************************ 00:09:45.956 START TEST raid_state_function_test_sb 00:09:45.956 ************************************ 00:09:45.956 19:00:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid0 4 true 00:09:45.956 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:45.956 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:45.957 Process raid pid: 80269 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=80269 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 80269' 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 80269 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 80269 ']' 00:09:45.957 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:45.957 19:00:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:45.957 [2024-12-05 19:00:03.433394] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:09:45.957 [2024-12-05 19:00:03.433535] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:46.217 [2024-12-05 19:00:03.589167] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:46.217 [2024-12-05 19:00:03.614140] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:46.217 [2024-12-05 19:00:03.657243] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:46.217 [2024-12-05 19:00:03.657283] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:46.787 [2024-12-05 19:00:04.256411] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:46.787 [2024-12-05 19:00:04.256479] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:46.787 [2024-12-05 19:00:04.256490] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:46.787 [2024-12-05 19:00:04.256585] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:46.787 [2024-12-05 19:00:04.256591] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:46.787 [2024-12-05 19:00:04.256603] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:46.787 [2024-12-05 19:00:04.256608] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:46.787 [2024-12-05 19:00:04.256617] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:46.787 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:46.788 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.788 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:46.788 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:46.788 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:46.788 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:46.788 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:46.788 "name": "Existed_Raid", 00:09:46.788 "uuid": "ee9e0377-8603-42dd-b596-2935159b47f5", 00:09:46.788 "strip_size_kb": 64, 00:09:46.788 "state": "configuring", 00:09:46.788 "raid_level": "raid0", 00:09:46.788 "superblock": true, 00:09:46.788 "num_base_bdevs": 4, 00:09:46.788 "num_base_bdevs_discovered": 0, 00:09:46.788 "num_base_bdevs_operational": 4, 00:09:46.788 "base_bdevs_list": [ 00:09:46.788 { 00:09:46.788 "name": "BaseBdev1", 00:09:46.788 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.788 "is_configured": false, 00:09:46.788 "data_offset": 0, 00:09:46.788 "data_size": 0 00:09:46.788 }, 00:09:46.788 { 00:09:46.788 "name": "BaseBdev2", 00:09:46.788 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.788 "is_configured": false, 00:09:46.788 "data_offset": 0, 00:09:46.788 "data_size": 0 00:09:46.788 }, 00:09:46.788 { 00:09:46.788 "name": "BaseBdev3", 00:09:46.788 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.788 "is_configured": false, 00:09:46.788 "data_offset": 0, 00:09:46.788 "data_size": 0 00:09:46.788 }, 00:09:46.788 { 00:09:46.788 "name": "BaseBdev4", 00:09:46.788 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:46.788 "is_configured": false, 00:09:46.788 "data_offset": 0, 00:09:46.788 "data_size": 0 00:09:46.788 } 00:09:46.788 ] 00:09:46.788 }' 00:09:46.788 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:46.788 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:47.359 [2024-12-05 19:00:04.659650] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:47.359 [2024-12-05 19:00:04.659698] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:47.359 [2024-12-05 19:00:04.671665] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:47.359 [2024-12-05 19:00:04.671719] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:47.359 [2024-12-05 19:00:04.671729] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:47.359 [2024-12-05 19:00:04.671760] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:47.359 [2024-12-05 19:00:04.671766] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:47.359 [2024-12-05 19:00:04.671775] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:47.359 [2024-12-05 19:00:04.671780] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:47.359 [2024-12-05 19:00:04.671789] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:47.359 [2024-12-05 19:00:04.692471] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:47.359 BaseBdev1 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:47.359 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:47.360 [ 00:09:47.360 { 00:09:47.360 "name": "BaseBdev1", 00:09:47.360 "aliases": [ 00:09:47.360 "5199d316-2bbf-4731-b20b-36cf9580489f" 00:09:47.360 ], 00:09:47.360 "product_name": "Malloc disk", 00:09:47.360 "block_size": 512, 00:09:47.360 "num_blocks": 65536, 00:09:47.360 "uuid": "5199d316-2bbf-4731-b20b-36cf9580489f", 00:09:47.360 "assigned_rate_limits": { 00:09:47.360 "rw_ios_per_sec": 0, 00:09:47.360 "rw_mbytes_per_sec": 0, 00:09:47.360 "r_mbytes_per_sec": 0, 00:09:47.360 "w_mbytes_per_sec": 0 00:09:47.360 }, 00:09:47.360 "claimed": true, 00:09:47.360 "claim_type": "exclusive_write", 00:09:47.360 "zoned": false, 00:09:47.360 "supported_io_types": { 00:09:47.360 "read": true, 00:09:47.360 "write": true, 00:09:47.360 "unmap": true, 00:09:47.360 "flush": true, 00:09:47.360 "reset": true, 00:09:47.360 "nvme_admin": false, 00:09:47.360 "nvme_io": false, 00:09:47.360 "nvme_io_md": false, 00:09:47.360 "write_zeroes": true, 00:09:47.360 "zcopy": true, 00:09:47.360 "get_zone_info": false, 00:09:47.360 "zone_management": false, 00:09:47.360 "zone_append": false, 00:09:47.360 "compare": false, 00:09:47.360 "compare_and_write": false, 00:09:47.360 "abort": true, 00:09:47.360 "seek_hole": false, 00:09:47.360 "seek_data": false, 00:09:47.360 "copy": true, 00:09:47.360 "nvme_iov_md": false 00:09:47.360 }, 00:09:47.360 "memory_domains": [ 00:09:47.360 { 00:09:47.360 "dma_device_id": "system", 00:09:47.360 "dma_device_type": 1 00:09:47.360 }, 00:09:47.360 { 00:09:47.360 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:47.360 "dma_device_type": 2 00:09:47.360 } 00:09:47.360 ], 00:09:47.360 "driver_specific": {} 00:09:47.360 } 00:09:47.360 ] 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:47.360 "name": "Existed_Raid", 00:09:47.360 "uuid": "51d55827-b364-4880-9252-54cacf06cbf4", 00:09:47.360 "strip_size_kb": 64, 00:09:47.360 "state": "configuring", 00:09:47.360 "raid_level": "raid0", 00:09:47.360 "superblock": true, 00:09:47.360 "num_base_bdevs": 4, 00:09:47.360 "num_base_bdevs_discovered": 1, 00:09:47.360 "num_base_bdevs_operational": 4, 00:09:47.360 "base_bdevs_list": [ 00:09:47.360 { 00:09:47.360 "name": "BaseBdev1", 00:09:47.360 "uuid": "5199d316-2bbf-4731-b20b-36cf9580489f", 00:09:47.360 "is_configured": true, 00:09:47.360 "data_offset": 2048, 00:09:47.360 "data_size": 63488 00:09:47.360 }, 00:09:47.360 { 00:09:47.360 "name": "BaseBdev2", 00:09:47.360 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:47.360 "is_configured": false, 00:09:47.360 "data_offset": 0, 00:09:47.360 "data_size": 0 00:09:47.360 }, 00:09:47.360 { 00:09:47.360 "name": "BaseBdev3", 00:09:47.360 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:47.360 "is_configured": false, 00:09:47.360 "data_offset": 0, 00:09:47.360 "data_size": 0 00:09:47.360 }, 00:09:47.360 { 00:09:47.360 "name": "BaseBdev4", 00:09:47.360 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:47.360 "is_configured": false, 00:09:47.360 "data_offset": 0, 00:09:47.360 "data_size": 0 00:09:47.360 } 00:09:47.360 ] 00:09:47.360 }' 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:47.360 19:00:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:47.629 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:47.629 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.629 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:47.629 [2024-12-05 19:00:05.155719] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:47.629 [2024-12-05 19:00:05.155816] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:47.629 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.629 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:47.629 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.629 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:47.629 [2024-12-05 19:00:05.167736] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:47.629 [2024-12-05 19:00:05.169606] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:47.629 [2024-12-05 19:00:05.169704] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:47.629 [2024-12-05 19:00:05.169737] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:47.629 [2024-12-05 19:00:05.169760] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:47.629 [2024-12-05 19:00:05.169779] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:47.629 [2024-12-05 19:00:05.169799] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:47.629 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:47.630 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:47.891 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:47.891 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:47.891 "name": "Existed_Raid", 00:09:47.891 "uuid": "7ff3ddc7-805c-4c3c-be77-e2209dd6c3f2", 00:09:47.891 "strip_size_kb": 64, 00:09:47.891 "state": "configuring", 00:09:47.891 "raid_level": "raid0", 00:09:47.891 "superblock": true, 00:09:47.891 "num_base_bdevs": 4, 00:09:47.891 "num_base_bdevs_discovered": 1, 00:09:47.891 "num_base_bdevs_operational": 4, 00:09:47.891 "base_bdevs_list": [ 00:09:47.891 { 00:09:47.891 "name": "BaseBdev1", 00:09:47.891 "uuid": "5199d316-2bbf-4731-b20b-36cf9580489f", 00:09:47.891 "is_configured": true, 00:09:47.891 "data_offset": 2048, 00:09:47.891 "data_size": 63488 00:09:47.891 }, 00:09:47.891 { 00:09:47.891 "name": "BaseBdev2", 00:09:47.891 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:47.891 "is_configured": false, 00:09:47.891 "data_offset": 0, 00:09:47.891 "data_size": 0 00:09:47.891 }, 00:09:47.891 { 00:09:47.891 "name": "BaseBdev3", 00:09:47.891 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:47.891 "is_configured": false, 00:09:47.891 "data_offset": 0, 00:09:47.891 "data_size": 0 00:09:47.891 }, 00:09:47.891 { 00:09:47.891 "name": "BaseBdev4", 00:09:47.891 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:47.891 "is_configured": false, 00:09:47.891 "data_offset": 0, 00:09:47.891 "data_size": 0 00:09:47.891 } 00:09:47.891 ] 00:09:47.891 }' 00:09:47.891 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:47.891 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:48.151 [2024-12-05 19:00:05.570024] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:48.151 BaseBdev2 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:48.151 [ 00:09:48.151 { 00:09:48.151 "name": "BaseBdev2", 00:09:48.151 "aliases": [ 00:09:48.151 "f3e3e4d1-a691-4d5b-afb8-1d2d457c4936" 00:09:48.151 ], 00:09:48.151 "product_name": "Malloc disk", 00:09:48.151 "block_size": 512, 00:09:48.151 "num_blocks": 65536, 00:09:48.151 "uuid": "f3e3e4d1-a691-4d5b-afb8-1d2d457c4936", 00:09:48.151 "assigned_rate_limits": { 00:09:48.151 "rw_ios_per_sec": 0, 00:09:48.151 "rw_mbytes_per_sec": 0, 00:09:48.151 "r_mbytes_per_sec": 0, 00:09:48.151 "w_mbytes_per_sec": 0 00:09:48.151 }, 00:09:48.151 "claimed": true, 00:09:48.151 "claim_type": "exclusive_write", 00:09:48.151 "zoned": false, 00:09:48.151 "supported_io_types": { 00:09:48.151 "read": true, 00:09:48.151 "write": true, 00:09:48.151 "unmap": true, 00:09:48.151 "flush": true, 00:09:48.151 "reset": true, 00:09:48.151 "nvme_admin": false, 00:09:48.151 "nvme_io": false, 00:09:48.151 "nvme_io_md": false, 00:09:48.151 "write_zeroes": true, 00:09:48.151 "zcopy": true, 00:09:48.151 "get_zone_info": false, 00:09:48.151 "zone_management": false, 00:09:48.151 "zone_append": false, 00:09:48.151 "compare": false, 00:09:48.151 "compare_and_write": false, 00:09:48.151 "abort": true, 00:09:48.151 "seek_hole": false, 00:09:48.151 "seek_data": false, 00:09:48.151 "copy": true, 00:09:48.151 "nvme_iov_md": false 00:09:48.151 }, 00:09:48.151 "memory_domains": [ 00:09:48.151 { 00:09:48.151 "dma_device_id": "system", 00:09:48.151 "dma_device_type": 1 00:09:48.151 }, 00:09:48.151 { 00:09:48.151 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:48.151 "dma_device_type": 2 00:09:48.151 } 00:09:48.151 ], 00:09:48.151 "driver_specific": {} 00:09:48.151 } 00:09:48.151 ] 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:48.151 "name": "Existed_Raid", 00:09:48.151 "uuid": "7ff3ddc7-805c-4c3c-be77-e2209dd6c3f2", 00:09:48.151 "strip_size_kb": 64, 00:09:48.151 "state": "configuring", 00:09:48.151 "raid_level": "raid0", 00:09:48.151 "superblock": true, 00:09:48.151 "num_base_bdevs": 4, 00:09:48.151 "num_base_bdevs_discovered": 2, 00:09:48.151 "num_base_bdevs_operational": 4, 00:09:48.151 "base_bdevs_list": [ 00:09:48.151 { 00:09:48.151 "name": "BaseBdev1", 00:09:48.151 "uuid": "5199d316-2bbf-4731-b20b-36cf9580489f", 00:09:48.151 "is_configured": true, 00:09:48.151 "data_offset": 2048, 00:09:48.151 "data_size": 63488 00:09:48.151 }, 00:09:48.151 { 00:09:48.151 "name": "BaseBdev2", 00:09:48.151 "uuid": "f3e3e4d1-a691-4d5b-afb8-1d2d457c4936", 00:09:48.151 "is_configured": true, 00:09:48.151 "data_offset": 2048, 00:09:48.151 "data_size": 63488 00:09:48.151 }, 00:09:48.151 { 00:09:48.151 "name": "BaseBdev3", 00:09:48.151 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.151 "is_configured": false, 00:09:48.151 "data_offset": 0, 00:09:48.151 "data_size": 0 00:09:48.151 }, 00:09:48.151 { 00:09:48.151 "name": "BaseBdev4", 00:09:48.151 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.151 "is_configured": false, 00:09:48.151 "data_offset": 0, 00:09:48.151 "data_size": 0 00:09:48.151 } 00:09:48.151 ] 00:09:48.151 }' 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:48.151 19:00:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:48.720 [2024-12-05 19:00:06.040266] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:48.720 BaseBdev3 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.720 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:48.720 [ 00:09:48.720 { 00:09:48.720 "name": "BaseBdev3", 00:09:48.720 "aliases": [ 00:09:48.720 "74575eae-94a6-48c5-8202-1c84bc6001ef" 00:09:48.720 ], 00:09:48.720 "product_name": "Malloc disk", 00:09:48.720 "block_size": 512, 00:09:48.721 "num_blocks": 65536, 00:09:48.721 "uuid": "74575eae-94a6-48c5-8202-1c84bc6001ef", 00:09:48.721 "assigned_rate_limits": { 00:09:48.721 "rw_ios_per_sec": 0, 00:09:48.721 "rw_mbytes_per_sec": 0, 00:09:48.721 "r_mbytes_per_sec": 0, 00:09:48.721 "w_mbytes_per_sec": 0 00:09:48.721 }, 00:09:48.721 "claimed": true, 00:09:48.721 "claim_type": "exclusive_write", 00:09:48.721 "zoned": false, 00:09:48.721 "supported_io_types": { 00:09:48.721 "read": true, 00:09:48.721 "write": true, 00:09:48.721 "unmap": true, 00:09:48.721 "flush": true, 00:09:48.721 "reset": true, 00:09:48.721 "nvme_admin": false, 00:09:48.721 "nvme_io": false, 00:09:48.721 "nvme_io_md": false, 00:09:48.721 "write_zeroes": true, 00:09:48.721 "zcopy": true, 00:09:48.721 "get_zone_info": false, 00:09:48.721 "zone_management": false, 00:09:48.721 "zone_append": false, 00:09:48.721 "compare": false, 00:09:48.721 "compare_and_write": false, 00:09:48.721 "abort": true, 00:09:48.721 "seek_hole": false, 00:09:48.721 "seek_data": false, 00:09:48.721 "copy": true, 00:09:48.721 "nvme_iov_md": false 00:09:48.721 }, 00:09:48.721 "memory_domains": [ 00:09:48.721 { 00:09:48.721 "dma_device_id": "system", 00:09:48.721 "dma_device_type": 1 00:09:48.721 }, 00:09:48.721 { 00:09:48.721 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:48.721 "dma_device_type": 2 00:09:48.721 } 00:09:48.721 ], 00:09:48.721 "driver_specific": {} 00:09:48.721 } 00:09:48.721 ] 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:48.721 "name": "Existed_Raid", 00:09:48.721 "uuid": "7ff3ddc7-805c-4c3c-be77-e2209dd6c3f2", 00:09:48.721 "strip_size_kb": 64, 00:09:48.721 "state": "configuring", 00:09:48.721 "raid_level": "raid0", 00:09:48.721 "superblock": true, 00:09:48.721 "num_base_bdevs": 4, 00:09:48.721 "num_base_bdevs_discovered": 3, 00:09:48.721 "num_base_bdevs_operational": 4, 00:09:48.721 "base_bdevs_list": [ 00:09:48.721 { 00:09:48.721 "name": "BaseBdev1", 00:09:48.721 "uuid": "5199d316-2bbf-4731-b20b-36cf9580489f", 00:09:48.721 "is_configured": true, 00:09:48.721 "data_offset": 2048, 00:09:48.721 "data_size": 63488 00:09:48.721 }, 00:09:48.721 { 00:09:48.721 "name": "BaseBdev2", 00:09:48.721 "uuid": "f3e3e4d1-a691-4d5b-afb8-1d2d457c4936", 00:09:48.721 "is_configured": true, 00:09:48.721 "data_offset": 2048, 00:09:48.721 "data_size": 63488 00:09:48.721 }, 00:09:48.721 { 00:09:48.721 "name": "BaseBdev3", 00:09:48.721 "uuid": "74575eae-94a6-48c5-8202-1c84bc6001ef", 00:09:48.721 "is_configured": true, 00:09:48.721 "data_offset": 2048, 00:09:48.721 "data_size": 63488 00:09:48.721 }, 00:09:48.721 { 00:09:48.721 "name": "BaseBdev4", 00:09:48.721 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.721 "is_configured": false, 00:09:48.721 "data_offset": 0, 00:09:48.721 "data_size": 0 00:09:48.721 } 00:09:48.721 ] 00:09:48.721 }' 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:48.721 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:48.981 [2024-12-05 19:00:06.498578] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:48.981 BaseBdev4 00:09:48.981 [2024-12-05 19:00:06.498876] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:48.981 [2024-12-05 19:00:06.498898] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:48.981 [2024-12-05 19:00:06.499202] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:48.981 [2024-12-05 19:00:06.499329] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:48.981 [2024-12-05 19:00:06.499342] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:48.981 [2024-12-05 19:00:06.499478] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:48.981 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:48.981 [ 00:09:48.981 { 00:09:48.981 "name": "BaseBdev4", 00:09:48.981 "aliases": [ 00:09:48.981 "28c7f359-155c-4a0b-8b71-e8cac4e84f6b" 00:09:48.981 ], 00:09:48.981 "product_name": "Malloc disk", 00:09:48.981 "block_size": 512, 00:09:48.981 "num_blocks": 65536, 00:09:48.981 "uuid": "28c7f359-155c-4a0b-8b71-e8cac4e84f6b", 00:09:48.981 "assigned_rate_limits": { 00:09:48.981 "rw_ios_per_sec": 0, 00:09:48.981 "rw_mbytes_per_sec": 0, 00:09:48.981 "r_mbytes_per_sec": 0, 00:09:48.981 "w_mbytes_per_sec": 0 00:09:48.981 }, 00:09:48.981 "claimed": true, 00:09:48.981 "claim_type": "exclusive_write", 00:09:48.981 "zoned": false, 00:09:48.981 "supported_io_types": { 00:09:48.981 "read": true, 00:09:48.981 "write": true, 00:09:48.981 "unmap": true, 00:09:48.981 "flush": true, 00:09:48.981 "reset": true, 00:09:48.981 "nvme_admin": false, 00:09:48.981 "nvme_io": false, 00:09:48.981 "nvme_io_md": false, 00:09:48.981 "write_zeroes": true, 00:09:48.981 "zcopy": true, 00:09:48.981 "get_zone_info": false, 00:09:48.981 "zone_management": false, 00:09:48.981 "zone_append": false, 00:09:48.981 "compare": false, 00:09:48.981 "compare_and_write": false, 00:09:48.981 "abort": true, 00:09:48.981 "seek_hole": false, 00:09:48.981 "seek_data": false, 00:09:48.981 "copy": true, 00:09:48.982 "nvme_iov_md": false 00:09:48.982 }, 00:09:48.982 "memory_domains": [ 00:09:48.982 { 00:09:48.982 "dma_device_id": "system", 00:09:48.982 "dma_device_type": 1 00:09:48.982 }, 00:09:48.982 { 00:09:48.982 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:48.982 "dma_device_type": 2 00:09:48.982 } 00:09:48.982 ], 00:09:48.982 "driver_specific": {} 00:09:48.982 } 00:09:48.982 ] 00:09:48.982 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:48.982 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:48.982 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:48.982 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:48.982 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:48.982 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:48.982 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:48.982 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:48.982 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:48.982 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:48.982 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.982 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.982 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.982 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:49.240 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.240 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.240 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.240 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:49.240 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.241 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:49.241 "name": "Existed_Raid", 00:09:49.241 "uuid": "7ff3ddc7-805c-4c3c-be77-e2209dd6c3f2", 00:09:49.241 "strip_size_kb": 64, 00:09:49.241 "state": "online", 00:09:49.241 "raid_level": "raid0", 00:09:49.241 "superblock": true, 00:09:49.241 "num_base_bdevs": 4, 00:09:49.241 "num_base_bdevs_discovered": 4, 00:09:49.241 "num_base_bdevs_operational": 4, 00:09:49.241 "base_bdevs_list": [ 00:09:49.241 { 00:09:49.241 "name": "BaseBdev1", 00:09:49.241 "uuid": "5199d316-2bbf-4731-b20b-36cf9580489f", 00:09:49.241 "is_configured": true, 00:09:49.241 "data_offset": 2048, 00:09:49.241 "data_size": 63488 00:09:49.241 }, 00:09:49.241 { 00:09:49.241 "name": "BaseBdev2", 00:09:49.241 "uuid": "f3e3e4d1-a691-4d5b-afb8-1d2d457c4936", 00:09:49.241 "is_configured": true, 00:09:49.241 "data_offset": 2048, 00:09:49.241 "data_size": 63488 00:09:49.241 }, 00:09:49.241 { 00:09:49.241 "name": "BaseBdev3", 00:09:49.241 "uuid": "74575eae-94a6-48c5-8202-1c84bc6001ef", 00:09:49.241 "is_configured": true, 00:09:49.241 "data_offset": 2048, 00:09:49.241 "data_size": 63488 00:09:49.241 }, 00:09:49.241 { 00:09:49.241 "name": "BaseBdev4", 00:09:49.241 "uuid": "28c7f359-155c-4a0b-8b71-e8cac4e84f6b", 00:09:49.241 "is_configured": true, 00:09:49.241 "data_offset": 2048, 00:09:49.241 "data_size": 63488 00:09:49.241 } 00:09:49.241 ] 00:09:49.241 }' 00:09:49.241 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:49.241 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.498 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:49.499 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:49.499 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:49.499 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:49.499 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:49.499 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:49.499 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:49.499 19:00:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:49.499 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.499 19:00:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.499 [2024-12-05 19:00:06.990073] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:49.499 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.499 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:49.499 "name": "Existed_Raid", 00:09:49.499 "aliases": [ 00:09:49.499 "7ff3ddc7-805c-4c3c-be77-e2209dd6c3f2" 00:09:49.499 ], 00:09:49.499 "product_name": "Raid Volume", 00:09:49.499 "block_size": 512, 00:09:49.499 "num_blocks": 253952, 00:09:49.499 "uuid": "7ff3ddc7-805c-4c3c-be77-e2209dd6c3f2", 00:09:49.499 "assigned_rate_limits": { 00:09:49.499 "rw_ios_per_sec": 0, 00:09:49.499 "rw_mbytes_per_sec": 0, 00:09:49.499 "r_mbytes_per_sec": 0, 00:09:49.499 "w_mbytes_per_sec": 0 00:09:49.499 }, 00:09:49.499 "claimed": false, 00:09:49.499 "zoned": false, 00:09:49.499 "supported_io_types": { 00:09:49.499 "read": true, 00:09:49.499 "write": true, 00:09:49.499 "unmap": true, 00:09:49.499 "flush": true, 00:09:49.499 "reset": true, 00:09:49.499 "nvme_admin": false, 00:09:49.499 "nvme_io": false, 00:09:49.499 "nvme_io_md": false, 00:09:49.499 "write_zeroes": true, 00:09:49.499 "zcopy": false, 00:09:49.499 "get_zone_info": false, 00:09:49.499 "zone_management": false, 00:09:49.499 "zone_append": false, 00:09:49.499 "compare": false, 00:09:49.499 "compare_and_write": false, 00:09:49.499 "abort": false, 00:09:49.499 "seek_hole": false, 00:09:49.499 "seek_data": false, 00:09:49.499 "copy": false, 00:09:49.499 "nvme_iov_md": false 00:09:49.499 }, 00:09:49.499 "memory_domains": [ 00:09:49.499 { 00:09:49.499 "dma_device_id": "system", 00:09:49.499 "dma_device_type": 1 00:09:49.499 }, 00:09:49.499 { 00:09:49.499 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:49.499 "dma_device_type": 2 00:09:49.499 }, 00:09:49.499 { 00:09:49.499 "dma_device_id": "system", 00:09:49.499 "dma_device_type": 1 00:09:49.499 }, 00:09:49.499 { 00:09:49.499 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:49.499 "dma_device_type": 2 00:09:49.499 }, 00:09:49.499 { 00:09:49.499 "dma_device_id": "system", 00:09:49.499 "dma_device_type": 1 00:09:49.499 }, 00:09:49.499 { 00:09:49.499 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:49.499 "dma_device_type": 2 00:09:49.499 }, 00:09:49.499 { 00:09:49.499 "dma_device_id": "system", 00:09:49.499 "dma_device_type": 1 00:09:49.499 }, 00:09:49.499 { 00:09:49.499 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:49.499 "dma_device_type": 2 00:09:49.499 } 00:09:49.499 ], 00:09:49.499 "driver_specific": { 00:09:49.499 "raid": { 00:09:49.499 "uuid": "7ff3ddc7-805c-4c3c-be77-e2209dd6c3f2", 00:09:49.499 "strip_size_kb": 64, 00:09:49.499 "state": "online", 00:09:49.499 "raid_level": "raid0", 00:09:49.499 "superblock": true, 00:09:49.499 "num_base_bdevs": 4, 00:09:49.499 "num_base_bdevs_discovered": 4, 00:09:49.499 "num_base_bdevs_operational": 4, 00:09:49.499 "base_bdevs_list": [ 00:09:49.499 { 00:09:49.499 "name": "BaseBdev1", 00:09:49.499 "uuid": "5199d316-2bbf-4731-b20b-36cf9580489f", 00:09:49.499 "is_configured": true, 00:09:49.499 "data_offset": 2048, 00:09:49.499 "data_size": 63488 00:09:49.499 }, 00:09:49.499 { 00:09:49.499 "name": "BaseBdev2", 00:09:49.499 "uuid": "f3e3e4d1-a691-4d5b-afb8-1d2d457c4936", 00:09:49.499 "is_configured": true, 00:09:49.499 "data_offset": 2048, 00:09:49.499 "data_size": 63488 00:09:49.499 }, 00:09:49.499 { 00:09:49.499 "name": "BaseBdev3", 00:09:49.499 "uuid": "74575eae-94a6-48c5-8202-1c84bc6001ef", 00:09:49.499 "is_configured": true, 00:09:49.499 "data_offset": 2048, 00:09:49.499 "data_size": 63488 00:09:49.499 }, 00:09:49.499 { 00:09:49.499 "name": "BaseBdev4", 00:09:49.499 "uuid": "28c7f359-155c-4a0b-8b71-e8cac4e84f6b", 00:09:49.499 "is_configured": true, 00:09:49.499 "data_offset": 2048, 00:09:49.499 "data_size": 63488 00:09:49.499 } 00:09:49.499 ] 00:09:49.499 } 00:09:49.499 } 00:09:49.499 }' 00:09:49.499 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:49.758 BaseBdev2 00:09:49.758 BaseBdev3 00:09:49.758 BaseBdev4' 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:49.758 [2024-12-05 19:00:07.293312] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:49.758 [2024-12-05 19:00:07.293340] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:49.758 [2024-12-05 19:00:07.293396] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.758 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:49.759 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.020 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.020 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:50.020 "name": "Existed_Raid", 00:09:50.020 "uuid": "7ff3ddc7-805c-4c3c-be77-e2209dd6c3f2", 00:09:50.020 "strip_size_kb": 64, 00:09:50.020 "state": "offline", 00:09:50.020 "raid_level": "raid0", 00:09:50.020 "superblock": true, 00:09:50.020 "num_base_bdevs": 4, 00:09:50.020 "num_base_bdevs_discovered": 3, 00:09:50.020 "num_base_bdevs_operational": 3, 00:09:50.020 "base_bdevs_list": [ 00:09:50.020 { 00:09:50.020 "name": null, 00:09:50.020 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:50.020 "is_configured": false, 00:09:50.020 "data_offset": 0, 00:09:50.020 "data_size": 63488 00:09:50.020 }, 00:09:50.020 { 00:09:50.020 "name": "BaseBdev2", 00:09:50.020 "uuid": "f3e3e4d1-a691-4d5b-afb8-1d2d457c4936", 00:09:50.020 "is_configured": true, 00:09:50.020 "data_offset": 2048, 00:09:50.020 "data_size": 63488 00:09:50.020 }, 00:09:50.020 { 00:09:50.020 "name": "BaseBdev3", 00:09:50.020 "uuid": "74575eae-94a6-48c5-8202-1c84bc6001ef", 00:09:50.020 "is_configured": true, 00:09:50.020 "data_offset": 2048, 00:09:50.020 "data_size": 63488 00:09:50.020 }, 00:09:50.020 { 00:09:50.020 "name": "BaseBdev4", 00:09:50.020 "uuid": "28c7f359-155c-4a0b-8b71-e8cac4e84f6b", 00:09:50.020 "is_configured": true, 00:09:50.020 "data_offset": 2048, 00:09:50.021 "data_size": 63488 00:09:50.021 } 00:09:50.021 ] 00:09:50.021 }' 00:09:50.021 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:50.021 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.284 [2024-12-05 19:00:07.727886] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.284 [2024-12-05 19:00:07.786997] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.284 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.284 [2024-12-05 19:00:07.842329] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:50.544 [2024-12-05 19:00:07.842418] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.544 BaseBdev2 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.544 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.544 [ 00:09:50.544 { 00:09:50.544 "name": "BaseBdev2", 00:09:50.544 "aliases": [ 00:09:50.544 "24fedc63-7603-4297-80b4-44a24ed77454" 00:09:50.544 ], 00:09:50.544 "product_name": "Malloc disk", 00:09:50.544 "block_size": 512, 00:09:50.544 "num_blocks": 65536, 00:09:50.544 "uuid": "24fedc63-7603-4297-80b4-44a24ed77454", 00:09:50.544 "assigned_rate_limits": { 00:09:50.544 "rw_ios_per_sec": 0, 00:09:50.544 "rw_mbytes_per_sec": 0, 00:09:50.544 "r_mbytes_per_sec": 0, 00:09:50.544 "w_mbytes_per_sec": 0 00:09:50.544 }, 00:09:50.544 "claimed": false, 00:09:50.544 "zoned": false, 00:09:50.544 "supported_io_types": { 00:09:50.544 "read": true, 00:09:50.544 "write": true, 00:09:50.544 "unmap": true, 00:09:50.544 "flush": true, 00:09:50.544 "reset": true, 00:09:50.544 "nvme_admin": false, 00:09:50.544 "nvme_io": false, 00:09:50.544 "nvme_io_md": false, 00:09:50.544 "write_zeroes": true, 00:09:50.544 "zcopy": true, 00:09:50.544 "get_zone_info": false, 00:09:50.544 "zone_management": false, 00:09:50.545 "zone_append": false, 00:09:50.545 "compare": false, 00:09:50.545 "compare_and_write": false, 00:09:50.545 "abort": true, 00:09:50.545 "seek_hole": false, 00:09:50.545 "seek_data": false, 00:09:50.545 "copy": true, 00:09:50.545 "nvme_iov_md": false 00:09:50.545 }, 00:09:50.545 "memory_domains": [ 00:09:50.545 { 00:09:50.545 "dma_device_id": "system", 00:09:50.545 "dma_device_type": 1 00:09:50.545 }, 00:09:50.545 { 00:09:50.545 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:50.545 "dma_device_type": 2 00:09:50.545 } 00:09:50.545 ], 00:09:50.545 "driver_specific": {} 00:09:50.545 } 00:09:50.545 ] 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.545 BaseBdev3 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.545 19:00:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.545 [ 00:09:50.545 { 00:09:50.545 "name": "BaseBdev3", 00:09:50.545 "aliases": [ 00:09:50.545 "a4236481-8cf5-480f-a48f-ef33cd9cea4f" 00:09:50.545 ], 00:09:50.545 "product_name": "Malloc disk", 00:09:50.545 "block_size": 512, 00:09:50.545 "num_blocks": 65536, 00:09:50.545 "uuid": "a4236481-8cf5-480f-a48f-ef33cd9cea4f", 00:09:50.545 "assigned_rate_limits": { 00:09:50.545 "rw_ios_per_sec": 0, 00:09:50.545 "rw_mbytes_per_sec": 0, 00:09:50.545 "r_mbytes_per_sec": 0, 00:09:50.545 "w_mbytes_per_sec": 0 00:09:50.545 }, 00:09:50.545 "claimed": false, 00:09:50.545 "zoned": false, 00:09:50.545 "supported_io_types": { 00:09:50.545 "read": true, 00:09:50.545 "write": true, 00:09:50.545 "unmap": true, 00:09:50.545 "flush": true, 00:09:50.545 "reset": true, 00:09:50.545 "nvme_admin": false, 00:09:50.545 "nvme_io": false, 00:09:50.545 "nvme_io_md": false, 00:09:50.545 "write_zeroes": true, 00:09:50.545 "zcopy": true, 00:09:50.545 "get_zone_info": false, 00:09:50.545 "zone_management": false, 00:09:50.545 "zone_append": false, 00:09:50.545 "compare": false, 00:09:50.545 "compare_and_write": false, 00:09:50.545 "abort": true, 00:09:50.545 "seek_hole": false, 00:09:50.545 "seek_data": false, 00:09:50.545 "copy": true, 00:09:50.545 "nvme_iov_md": false 00:09:50.545 }, 00:09:50.545 "memory_domains": [ 00:09:50.545 { 00:09:50.545 "dma_device_id": "system", 00:09:50.545 "dma_device_type": 1 00:09:50.545 }, 00:09:50.545 { 00:09:50.545 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:50.545 "dma_device_type": 2 00:09:50.545 } 00:09:50.545 ], 00:09:50.545 "driver_specific": {} 00:09:50.545 } 00:09:50.545 ] 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.545 BaseBdev4 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.545 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.545 [ 00:09:50.545 { 00:09:50.545 "name": "BaseBdev4", 00:09:50.545 "aliases": [ 00:09:50.545 "24c8aad8-e79f-4681-9a0f-ba205e54643a" 00:09:50.545 ], 00:09:50.545 "product_name": "Malloc disk", 00:09:50.545 "block_size": 512, 00:09:50.545 "num_blocks": 65536, 00:09:50.545 "uuid": "24c8aad8-e79f-4681-9a0f-ba205e54643a", 00:09:50.545 "assigned_rate_limits": { 00:09:50.545 "rw_ios_per_sec": 0, 00:09:50.545 "rw_mbytes_per_sec": 0, 00:09:50.545 "r_mbytes_per_sec": 0, 00:09:50.545 "w_mbytes_per_sec": 0 00:09:50.545 }, 00:09:50.545 "claimed": false, 00:09:50.545 "zoned": false, 00:09:50.545 "supported_io_types": { 00:09:50.545 "read": true, 00:09:50.545 "write": true, 00:09:50.545 "unmap": true, 00:09:50.545 "flush": true, 00:09:50.545 "reset": true, 00:09:50.545 "nvme_admin": false, 00:09:50.545 "nvme_io": false, 00:09:50.545 "nvme_io_md": false, 00:09:50.545 "write_zeroes": true, 00:09:50.545 "zcopy": true, 00:09:50.545 "get_zone_info": false, 00:09:50.545 "zone_management": false, 00:09:50.545 "zone_append": false, 00:09:50.545 "compare": false, 00:09:50.545 "compare_and_write": false, 00:09:50.545 "abort": true, 00:09:50.545 "seek_hole": false, 00:09:50.545 "seek_data": false, 00:09:50.545 "copy": true, 00:09:50.545 "nvme_iov_md": false 00:09:50.545 }, 00:09:50.545 "memory_domains": [ 00:09:50.545 { 00:09:50.545 "dma_device_id": "system", 00:09:50.545 "dma_device_type": 1 00:09:50.545 }, 00:09:50.545 { 00:09:50.545 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:50.545 "dma_device_type": 2 00:09:50.545 } 00:09:50.545 ], 00:09:50.545 "driver_specific": {} 00:09:50.545 } 00:09:50.546 ] 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.546 [2024-12-05 19:00:08.070077] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:50.546 [2024-12-05 19:00:08.070161] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:50.546 [2024-12-05 19:00:08.070221] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:50.546 [2024-12-05 19:00:08.072074] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:50.546 [2024-12-05 19:00:08.072164] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:50.546 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:50.805 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:50.805 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:50.805 "name": "Existed_Raid", 00:09:50.805 "uuid": "544776e7-3805-4d41-91a0-6cbe41d84a6d", 00:09:50.805 "strip_size_kb": 64, 00:09:50.805 "state": "configuring", 00:09:50.805 "raid_level": "raid0", 00:09:50.805 "superblock": true, 00:09:50.805 "num_base_bdevs": 4, 00:09:50.805 "num_base_bdevs_discovered": 3, 00:09:50.805 "num_base_bdevs_operational": 4, 00:09:50.805 "base_bdevs_list": [ 00:09:50.805 { 00:09:50.805 "name": "BaseBdev1", 00:09:50.805 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:50.805 "is_configured": false, 00:09:50.805 "data_offset": 0, 00:09:50.805 "data_size": 0 00:09:50.805 }, 00:09:50.805 { 00:09:50.805 "name": "BaseBdev2", 00:09:50.805 "uuid": "24fedc63-7603-4297-80b4-44a24ed77454", 00:09:50.805 "is_configured": true, 00:09:50.805 "data_offset": 2048, 00:09:50.805 "data_size": 63488 00:09:50.805 }, 00:09:50.805 { 00:09:50.805 "name": "BaseBdev3", 00:09:50.805 "uuid": "a4236481-8cf5-480f-a48f-ef33cd9cea4f", 00:09:50.805 "is_configured": true, 00:09:50.805 "data_offset": 2048, 00:09:50.805 "data_size": 63488 00:09:50.805 }, 00:09:50.805 { 00:09:50.805 "name": "BaseBdev4", 00:09:50.805 "uuid": "24c8aad8-e79f-4681-9a0f-ba205e54643a", 00:09:50.805 "is_configured": true, 00:09:50.805 "data_offset": 2048, 00:09:50.805 "data_size": 63488 00:09:50.805 } 00:09:50.805 ] 00:09:50.805 }' 00:09:50.805 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:50.805 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.063 [2024-12-05 19:00:08.525292] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:51.063 "name": "Existed_Raid", 00:09:51.063 "uuid": "544776e7-3805-4d41-91a0-6cbe41d84a6d", 00:09:51.063 "strip_size_kb": 64, 00:09:51.063 "state": "configuring", 00:09:51.063 "raid_level": "raid0", 00:09:51.063 "superblock": true, 00:09:51.063 "num_base_bdevs": 4, 00:09:51.063 "num_base_bdevs_discovered": 2, 00:09:51.063 "num_base_bdevs_operational": 4, 00:09:51.063 "base_bdevs_list": [ 00:09:51.063 { 00:09:51.063 "name": "BaseBdev1", 00:09:51.063 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.063 "is_configured": false, 00:09:51.063 "data_offset": 0, 00:09:51.063 "data_size": 0 00:09:51.063 }, 00:09:51.063 { 00:09:51.063 "name": null, 00:09:51.063 "uuid": "24fedc63-7603-4297-80b4-44a24ed77454", 00:09:51.063 "is_configured": false, 00:09:51.063 "data_offset": 0, 00:09:51.063 "data_size": 63488 00:09:51.063 }, 00:09:51.063 { 00:09:51.063 "name": "BaseBdev3", 00:09:51.063 "uuid": "a4236481-8cf5-480f-a48f-ef33cd9cea4f", 00:09:51.063 "is_configured": true, 00:09:51.063 "data_offset": 2048, 00:09:51.063 "data_size": 63488 00:09:51.063 }, 00:09:51.063 { 00:09:51.063 "name": "BaseBdev4", 00:09:51.063 "uuid": "24c8aad8-e79f-4681-9a0f-ba205e54643a", 00:09:51.063 "is_configured": true, 00:09:51.063 "data_offset": 2048, 00:09:51.063 "data_size": 63488 00:09:51.063 } 00:09:51.063 ] 00:09:51.063 }' 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:51.063 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.631 [2024-12-05 19:00:08.951633] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:51.631 BaseBdev1 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.631 [ 00:09:51.631 { 00:09:51.631 "name": "BaseBdev1", 00:09:51.631 "aliases": [ 00:09:51.631 "4b90a385-b9bb-4891-8488-49508387782e" 00:09:51.631 ], 00:09:51.631 "product_name": "Malloc disk", 00:09:51.631 "block_size": 512, 00:09:51.631 "num_blocks": 65536, 00:09:51.631 "uuid": "4b90a385-b9bb-4891-8488-49508387782e", 00:09:51.631 "assigned_rate_limits": { 00:09:51.631 "rw_ios_per_sec": 0, 00:09:51.631 "rw_mbytes_per_sec": 0, 00:09:51.631 "r_mbytes_per_sec": 0, 00:09:51.631 "w_mbytes_per_sec": 0 00:09:51.631 }, 00:09:51.631 "claimed": true, 00:09:51.631 "claim_type": "exclusive_write", 00:09:51.631 "zoned": false, 00:09:51.631 "supported_io_types": { 00:09:51.631 "read": true, 00:09:51.631 "write": true, 00:09:51.631 "unmap": true, 00:09:51.631 "flush": true, 00:09:51.631 "reset": true, 00:09:51.631 "nvme_admin": false, 00:09:51.631 "nvme_io": false, 00:09:51.631 "nvme_io_md": false, 00:09:51.631 "write_zeroes": true, 00:09:51.631 "zcopy": true, 00:09:51.631 "get_zone_info": false, 00:09:51.631 "zone_management": false, 00:09:51.631 "zone_append": false, 00:09:51.631 "compare": false, 00:09:51.631 "compare_and_write": false, 00:09:51.631 "abort": true, 00:09:51.631 "seek_hole": false, 00:09:51.631 "seek_data": false, 00:09:51.631 "copy": true, 00:09:51.631 "nvme_iov_md": false 00:09:51.631 }, 00:09:51.631 "memory_domains": [ 00:09:51.631 { 00:09:51.631 "dma_device_id": "system", 00:09:51.631 "dma_device_type": 1 00:09:51.631 }, 00:09:51.631 { 00:09:51.631 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:51.631 "dma_device_type": 2 00:09:51.631 } 00:09:51.631 ], 00:09:51.631 "driver_specific": {} 00:09:51.631 } 00:09:51.631 ] 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:51.631 19:00:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:51.631 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:51.631 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:51.631 "name": "Existed_Raid", 00:09:51.631 "uuid": "544776e7-3805-4d41-91a0-6cbe41d84a6d", 00:09:51.631 "strip_size_kb": 64, 00:09:51.631 "state": "configuring", 00:09:51.631 "raid_level": "raid0", 00:09:51.631 "superblock": true, 00:09:51.631 "num_base_bdevs": 4, 00:09:51.631 "num_base_bdevs_discovered": 3, 00:09:51.631 "num_base_bdevs_operational": 4, 00:09:51.631 "base_bdevs_list": [ 00:09:51.631 { 00:09:51.631 "name": "BaseBdev1", 00:09:51.632 "uuid": "4b90a385-b9bb-4891-8488-49508387782e", 00:09:51.632 "is_configured": true, 00:09:51.632 "data_offset": 2048, 00:09:51.632 "data_size": 63488 00:09:51.632 }, 00:09:51.632 { 00:09:51.632 "name": null, 00:09:51.632 "uuid": "24fedc63-7603-4297-80b4-44a24ed77454", 00:09:51.632 "is_configured": false, 00:09:51.632 "data_offset": 0, 00:09:51.632 "data_size": 63488 00:09:51.632 }, 00:09:51.632 { 00:09:51.632 "name": "BaseBdev3", 00:09:51.632 "uuid": "a4236481-8cf5-480f-a48f-ef33cd9cea4f", 00:09:51.632 "is_configured": true, 00:09:51.632 "data_offset": 2048, 00:09:51.632 "data_size": 63488 00:09:51.632 }, 00:09:51.632 { 00:09:51.632 "name": "BaseBdev4", 00:09:51.632 "uuid": "24c8aad8-e79f-4681-9a0f-ba205e54643a", 00:09:51.632 "is_configured": true, 00:09:51.632 "data_offset": 2048, 00:09:51.632 "data_size": 63488 00:09:51.632 } 00:09:51.632 ] 00:09:51.632 }' 00:09:51.632 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:51.632 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.201 [2024-12-05 19:00:09.490770] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.201 "name": "Existed_Raid", 00:09:52.201 "uuid": "544776e7-3805-4d41-91a0-6cbe41d84a6d", 00:09:52.201 "strip_size_kb": 64, 00:09:52.201 "state": "configuring", 00:09:52.201 "raid_level": "raid0", 00:09:52.201 "superblock": true, 00:09:52.201 "num_base_bdevs": 4, 00:09:52.201 "num_base_bdevs_discovered": 2, 00:09:52.201 "num_base_bdevs_operational": 4, 00:09:52.201 "base_bdevs_list": [ 00:09:52.201 { 00:09:52.201 "name": "BaseBdev1", 00:09:52.201 "uuid": "4b90a385-b9bb-4891-8488-49508387782e", 00:09:52.201 "is_configured": true, 00:09:52.201 "data_offset": 2048, 00:09:52.201 "data_size": 63488 00:09:52.201 }, 00:09:52.201 { 00:09:52.201 "name": null, 00:09:52.201 "uuid": "24fedc63-7603-4297-80b4-44a24ed77454", 00:09:52.201 "is_configured": false, 00:09:52.201 "data_offset": 0, 00:09:52.201 "data_size": 63488 00:09:52.201 }, 00:09:52.201 { 00:09:52.201 "name": null, 00:09:52.201 "uuid": "a4236481-8cf5-480f-a48f-ef33cd9cea4f", 00:09:52.201 "is_configured": false, 00:09:52.201 "data_offset": 0, 00:09:52.201 "data_size": 63488 00:09:52.201 }, 00:09:52.201 { 00:09:52.201 "name": "BaseBdev4", 00:09:52.201 "uuid": "24c8aad8-e79f-4681-9a0f-ba205e54643a", 00:09:52.201 "is_configured": true, 00:09:52.201 "data_offset": 2048, 00:09:52.201 "data_size": 63488 00:09:52.201 } 00:09:52.201 ] 00:09:52.201 }' 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.201 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.461 [2024-12-05 19:00:09.953983] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:52.461 19:00:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.461 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.461 "name": "Existed_Raid", 00:09:52.461 "uuid": "544776e7-3805-4d41-91a0-6cbe41d84a6d", 00:09:52.461 "strip_size_kb": 64, 00:09:52.461 "state": "configuring", 00:09:52.461 "raid_level": "raid0", 00:09:52.461 "superblock": true, 00:09:52.461 "num_base_bdevs": 4, 00:09:52.461 "num_base_bdevs_discovered": 3, 00:09:52.461 "num_base_bdevs_operational": 4, 00:09:52.461 "base_bdevs_list": [ 00:09:52.461 { 00:09:52.461 "name": "BaseBdev1", 00:09:52.461 "uuid": "4b90a385-b9bb-4891-8488-49508387782e", 00:09:52.461 "is_configured": true, 00:09:52.461 "data_offset": 2048, 00:09:52.461 "data_size": 63488 00:09:52.461 }, 00:09:52.461 { 00:09:52.461 "name": null, 00:09:52.461 "uuid": "24fedc63-7603-4297-80b4-44a24ed77454", 00:09:52.461 "is_configured": false, 00:09:52.461 "data_offset": 0, 00:09:52.461 "data_size": 63488 00:09:52.461 }, 00:09:52.461 { 00:09:52.461 "name": "BaseBdev3", 00:09:52.461 "uuid": "a4236481-8cf5-480f-a48f-ef33cd9cea4f", 00:09:52.461 "is_configured": true, 00:09:52.461 "data_offset": 2048, 00:09:52.461 "data_size": 63488 00:09:52.461 }, 00:09:52.461 { 00:09:52.461 "name": "BaseBdev4", 00:09:52.461 "uuid": "24c8aad8-e79f-4681-9a0f-ba205e54643a", 00:09:52.461 "is_configured": true, 00:09:52.461 "data_offset": 2048, 00:09:52.461 "data_size": 63488 00:09:52.461 } 00:09:52.461 ] 00:09:52.461 }' 00:09:52.461 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.461 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.031 [2024-12-05 19:00:10.445173] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:53.031 "name": "Existed_Raid", 00:09:53.031 "uuid": "544776e7-3805-4d41-91a0-6cbe41d84a6d", 00:09:53.031 "strip_size_kb": 64, 00:09:53.031 "state": "configuring", 00:09:53.031 "raid_level": "raid0", 00:09:53.031 "superblock": true, 00:09:53.031 "num_base_bdevs": 4, 00:09:53.031 "num_base_bdevs_discovered": 2, 00:09:53.031 "num_base_bdevs_operational": 4, 00:09:53.031 "base_bdevs_list": [ 00:09:53.031 { 00:09:53.031 "name": null, 00:09:53.031 "uuid": "4b90a385-b9bb-4891-8488-49508387782e", 00:09:53.031 "is_configured": false, 00:09:53.031 "data_offset": 0, 00:09:53.031 "data_size": 63488 00:09:53.031 }, 00:09:53.031 { 00:09:53.031 "name": null, 00:09:53.031 "uuid": "24fedc63-7603-4297-80b4-44a24ed77454", 00:09:53.031 "is_configured": false, 00:09:53.031 "data_offset": 0, 00:09:53.031 "data_size": 63488 00:09:53.031 }, 00:09:53.031 { 00:09:53.031 "name": "BaseBdev3", 00:09:53.031 "uuid": "a4236481-8cf5-480f-a48f-ef33cd9cea4f", 00:09:53.031 "is_configured": true, 00:09:53.031 "data_offset": 2048, 00:09:53.031 "data_size": 63488 00:09:53.031 }, 00:09:53.031 { 00:09:53.031 "name": "BaseBdev4", 00:09:53.031 "uuid": "24c8aad8-e79f-4681-9a0f-ba205e54643a", 00:09:53.031 "is_configured": true, 00:09:53.031 "data_offset": 2048, 00:09:53.031 "data_size": 63488 00:09:53.031 } 00:09:53.031 ] 00:09:53.031 }' 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:53.031 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.602 [2024-12-05 19:00:10.950796] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.602 19:00:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:53.602 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:53.602 "name": "Existed_Raid", 00:09:53.602 "uuid": "544776e7-3805-4d41-91a0-6cbe41d84a6d", 00:09:53.602 "strip_size_kb": 64, 00:09:53.602 "state": "configuring", 00:09:53.602 "raid_level": "raid0", 00:09:53.602 "superblock": true, 00:09:53.602 "num_base_bdevs": 4, 00:09:53.602 "num_base_bdevs_discovered": 3, 00:09:53.602 "num_base_bdevs_operational": 4, 00:09:53.602 "base_bdevs_list": [ 00:09:53.602 { 00:09:53.602 "name": null, 00:09:53.602 "uuid": "4b90a385-b9bb-4891-8488-49508387782e", 00:09:53.602 "is_configured": false, 00:09:53.602 "data_offset": 0, 00:09:53.602 "data_size": 63488 00:09:53.602 }, 00:09:53.602 { 00:09:53.602 "name": "BaseBdev2", 00:09:53.602 "uuid": "24fedc63-7603-4297-80b4-44a24ed77454", 00:09:53.602 "is_configured": true, 00:09:53.602 "data_offset": 2048, 00:09:53.602 "data_size": 63488 00:09:53.602 }, 00:09:53.602 { 00:09:53.602 "name": "BaseBdev3", 00:09:53.602 "uuid": "a4236481-8cf5-480f-a48f-ef33cd9cea4f", 00:09:53.602 "is_configured": true, 00:09:53.602 "data_offset": 2048, 00:09:53.602 "data_size": 63488 00:09:53.602 }, 00:09:53.602 { 00:09:53.602 "name": "BaseBdev4", 00:09:53.602 "uuid": "24c8aad8-e79f-4681-9a0f-ba205e54643a", 00:09:53.602 "is_configured": true, 00:09:53.602 "data_offset": 2048, 00:09:53.602 "data_size": 63488 00:09:53.602 } 00:09:53.602 ] 00:09:53.602 }' 00:09:53.602 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:53.602 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.862 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.862 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:53.862 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:53.862 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:53.862 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 4b90a385-b9bb-4891-8488-49508387782e 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.123 [2024-12-05 19:00:11.484849] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:54.123 [2024-12-05 19:00:11.485126] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:54.123 [2024-12-05 19:00:11.485144] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:54.123 [2024-12-05 19:00:11.485398] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:09:54.123 [2024-12-05 19:00:11.485509] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:54.123 [2024-12-05 19:00:11.485519] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:54.123 [2024-12-05 19:00:11.485610] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:54.123 NewBaseBdev 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.123 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.123 [ 00:09:54.123 { 00:09:54.123 "name": "NewBaseBdev", 00:09:54.123 "aliases": [ 00:09:54.123 "4b90a385-b9bb-4891-8488-49508387782e" 00:09:54.123 ], 00:09:54.123 "product_name": "Malloc disk", 00:09:54.123 "block_size": 512, 00:09:54.123 "num_blocks": 65536, 00:09:54.123 "uuid": "4b90a385-b9bb-4891-8488-49508387782e", 00:09:54.123 "assigned_rate_limits": { 00:09:54.123 "rw_ios_per_sec": 0, 00:09:54.123 "rw_mbytes_per_sec": 0, 00:09:54.123 "r_mbytes_per_sec": 0, 00:09:54.123 "w_mbytes_per_sec": 0 00:09:54.123 }, 00:09:54.123 "claimed": true, 00:09:54.123 "claim_type": "exclusive_write", 00:09:54.123 "zoned": false, 00:09:54.123 "supported_io_types": { 00:09:54.123 "read": true, 00:09:54.123 "write": true, 00:09:54.123 "unmap": true, 00:09:54.123 "flush": true, 00:09:54.123 "reset": true, 00:09:54.123 "nvme_admin": false, 00:09:54.123 "nvme_io": false, 00:09:54.123 "nvme_io_md": false, 00:09:54.123 "write_zeroes": true, 00:09:54.123 "zcopy": true, 00:09:54.123 "get_zone_info": false, 00:09:54.123 "zone_management": false, 00:09:54.123 "zone_append": false, 00:09:54.123 "compare": false, 00:09:54.123 "compare_and_write": false, 00:09:54.123 "abort": true, 00:09:54.123 "seek_hole": false, 00:09:54.123 "seek_data": false, 00:09:54.123 "copy": true, 00:09:54.123 "nvme_iov_md": false 00:09:54.123 }, 00:09:54.123 "memory_domains": [ 00:09:54.123 { 00:09:54.123 "dma_device_id": "system", 00:09:54.123 "dma_device_type": 1 00:09:54.124 }, 00:09:54.124 { 00:09:54.124 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:54.124 "dma_device_type": 2 00:09:54.124 } 00:09:54.124 ], 00:09:54.124 "driver_specific": {} 00:09:54.124 } 00:09:54.124 ] 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:54.124 "name": "Existed_Raid", 00:09:54.124 "uuid": "544776e7-3805-4d41-91a0-6cbe41d84a6d", 00:09:54.124 "strip_size_kb": 64, 00:09:54.124 "state": "online", 00:09:54.124 "raid_level": "raid0", 00:09:54.124 "superblock": true, 00:09:54.124 "num_base_bdevs": 4, 00:09:54.124 "num_base_bdevs_discovered": 4, 00:09:54.124 "num_base_bdevs_operational": 4, 00:09:54.124 "base_bdevs_list": [ 00:09:54.124 { 00:09:54.124 "name": "NewBaseBdev", 00:09:54.124 "uuid": "4b90a385-b9bb-4891-8488-49508387782e", 00:09:54.124 "is_configured": true, 00:09:54.124 "data_offset": 2048, 00:09:54.124 "data_size": 63488 00:09:54.124 }, 00:09:54.124 { 00:09:54.124 "name": "BaseBdev2", 00:09:54.124 "uuid": "24fedc63-7603-4297-80b4-44a24ed77454", 00:09:54.124 "is_configured": true, 00:09:54.124 "data_offset": 2048, 00:09:54.124 "data_size": 63488 00:09:54.124 }, 00:09:54.124 { 00:09:54.124 "name": "BaseBdev3", 00:09:54.124 "uuid": "a4236481-8cf5-480f-a48f-ef33cd9cea4f", 00:09:54.124 "is_configured": true, 00:09:54.124 "data_offset": 2048, 00:09:54.124 "data_size": 63488 00:09:54.124 }, 00:09:54.124 { 00:09:54.124 "name": "BaseBdev4", 00:09:54.124 "uuid": "24c8aad8-e79f-4681-9a0f-ba205e54643a", 00:09:54.124 "is_configured": true, 00:09:54.124 "data_offset": 2048, 00:09:54.124 "data_size": 63488 00:09:54.124 } 00:09:54.124 ] 00:09:54.124 }' 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:54.124 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.384 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:54.384 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:54.384 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:54.384 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:54.384 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:54.384 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:54.384 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:54.384 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.384 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.384 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:54.384 [2024-12-05 19:00:11.936396] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:54.644 19:00:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.644 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:54.644 "name": "Existed_Raid", 00:09:54.644 "aliases": [ 00:09:54.644 "544776e7-3805-4d41-91a0-6cbe41d84a6d" 00:09:54.644 ], 00:09:54.644 "product_name": "Raid Volume", 00:09:54.644 "block_size": 512, 00:09:54.644 "num_blocks": 253952, 00:09:54.644 "uuid": "544776e7-3805-4d41-91a0-6cbe41d84a6d", 00:09:54.644 "assigned_rate_limits": { 00:09:54.644 "rw_ios_per_sec": 0, 00:09:54.644 "rw_mbytes_per_sec": 0, 00:09:54.644 "r_mbytes_per_sec": 0, 00:09:54.644 "w_mbytes_per_sec": 0 00:09:54.644 }, 00:09:54.644 "claimed": false, 00:09:54.644 "zoned": false, 00:09:54.644 "supported_io_types": { 00:09:54.644 "read": true, 00:09:54.644 "write": true, 00:09:54.644 "unmap": true, 00:09:54.644 "flush": true, 00:09:54.644 "reset": true, 00:09:54.644 "nvme_admin": false, 00:09:54.644 "nvme_io": false, 00:09:54.644 "nvme_io_md": false, 00:09:54.644 "write_zeroes": true, 00:09:54.644 "zcopy": false, 00:09:54.644 "get_zone_info": false, 00:09:54.644 "zone_management": false, 00:09:54.644 "zone_append": false, 00:09:54.644 "compare": false, 00:09:54.644 "compare_and_write": false, 00:09:54.644 "abort": false, 00:09:54.644 "seek_hole": false, 00:09:54.644 "seek_data": false, 00:09:54.644 "copy": false, 00:09:54.644 "nvme_iov_md": false 00:09:54.644 }, 00:09:54.644 "memory_domains": [ 00:09:54.644 { 00:09:54.644 "dma_device_id": "system", 00:09:54.644 "dma_device_type": 1 00:09:54.644 }, 00:09:54.644 { 00:09:54.644 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:54.644 "dma_device_type": 2 00:09:54.644 }, 00:09:54.644 { 00:09:54.644 "dma_device_id": "system", 00:09:54.644 "dma_device_type": 1 00:09:54.644 }, 00:09:54.644 { 00:09:54.644 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:54.644 "dma_device_type": 2 00:09:54.644 }, 00:09:54.644 { 00:09:54.644 "dma_device_id": "system", 00:09:54.644 "dma_device_type": 1 00:09:54.644 }, 00:09:54.644 { 00:09:54.644 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:54.644 "dma_device_type": 2 00:09:54.644 }, 00:09:54.644 { 00:09:54.644 "dma_device_id": "system", 00:09:54.644 "dma_device_type": 1 00:09:54.644 }, 00:09:54.644 { 00:09:54.644 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:54.644 "dma_device_type": 2 00:09:54.644 } 00:09:54.644 ], 00:09:54.644 "driver_specific": { 00:09:54.644 "raid": { 00:09:54.644 "uuid": "544776e7-3805-4d41-91a0-6cbe41d84a6d", 00:09:54.644 "strip_size_kb": 64, 00:09:54.644 "state": "online", 00:09:54.644 "raid_level": "raid0", 00:09:54.644 "superblock": true, 00:09:54.644 "num_base_bdevs": 4, 00:09:54.644 "num_base_bdevs_discovered": 4, 00:09:54.644 "num_base_bdevs_operational": 4, 00:09:54.644 "base_bdevs_list": [ 00:09:54.644 { 00:09:54.644 "name": "NewBaseBdev", 00:09:54.644 "uuid": "4b90a385-b9bb-4891-8488-49508387782e", 00:09:54.644 "is_configured": true, 00:09:54.644 "data_offset": 2048, 00:09:54.644 "data_size": 63488 00:09:54.644 }, 00:09:54.644 { 00:09:54.644 "name": "BaseBdev2", 00:09:54.644 "uuid": "24fedc63-7603-4297-80b4-44a24ed77454", 00:09:54.644 "is_configured": true, 00:09:54.644 "data_offset": 2048, 00:09:54.645 "data_size": 63488 00:09:54.645 }, 00:09:54.645 { 00:09:54.645 "name": "BaseBdev3", 00:09:54.645 "uuid": "a4236481-8cf5-480f-a48f-ef33cd9cea4f", 00:09:54.645 "is_configured": true, 00:09:54.645 "data_offset": 2048, 00:09:54.645 "data_size": 63488 00:09:54.645 }, 00:09:54.645 { 00:09:54.645 "name": "BaseBdev4", 00:09:54.645 "uuid": "24c8aad8-e79f-4681-9a0f-ba205e54643a", 00:09:54.645 "is_configured": true, 00:09:54.645 "data_offset": 2048, 00:09:54.645 "data_size": 63488 00:09:54.645 } 00:09:54.645 ] 00:09:54.645 } 00:09:54.645 } 00:09:54.645 }' 00:09:54.645 19:00:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:54.645 BaseBdev2 00:09:54.645 BaseBdev3 00:09:54.645 BaseBdev4' 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.645 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:54.905 [2024-12-05 19:00:12.243726] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:54.905 [2024-12-05 19:00:12.243761] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:54.905 [2024-12-05 19:00:12.243822] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:54.905 [2024-12-05 19:00:12.243882] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:54.905 [2024-12-05 19:00:12.243891] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 80269 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 80269 ']' 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 80269 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 80269 00:09:54.905 killing process with pid 80269 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 80269' 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 80269 00:09:54.905 [2024-12-05 19:00:12.293950] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:54.905 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 80269 00:09:54.905 [2024-12-05 19:00:12.334418] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:55.165 ************************************ 00:09:55.165 END TEST raid_state_function_test_sb 00:09:55.165 ************************************ 00:09:55.165 19:00:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:55.165 00:09:55.165 real 0m9.212s 00:09:55.165 user 0m15.745s 00:09:55.165 sys 0m1.865s 00:09:55.165 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:55.165 19:00:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:55.165 19:00:12 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 4 00:09:55.165 19:00:12 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:09:55.165 19:00:12 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:55.165 19:00:12 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:55.165 ************************************ 00:09:55.165 START TEST raid_superblock_test 00:09:55.165 ************************************ 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid0 4 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=80863 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 80863 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 80863 ']' 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:55.165 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:55.165 19:00:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.165 [2024-12-05 19:00:12.708862] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:09:55.165 [2024-12-05 19:00:12.709093] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid80863 ] 00:09:55.426 [2024-12-05 19:00:12.863733] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:55.426 [2024-12-05 19:00:12.888208] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:55.426 [2024-12-05 19:00:12.931368] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:55.426 [2024-12-05 19:00:12.931496] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.996 malloc1 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:55.996 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.257 [2024-12-05 19:00:13.555932] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:56.257 [2024-12-05 19:00:13.556060] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:56.257 [2024-12-05 19:00:13.556107] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:09:56.257 [2024-12-05 19:00:13.556149] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:56.257 [2024-12-05 19:00:13.558287] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:56.257 [2024-12-05 19:00:13.558377] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:56.257 pt1 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.257 malloc2 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.257 [2024-12-05 19:00:13.588501] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:56.257 [2024-12-05 19:00:13.588559] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:56.257 [2024-12-05 19:00:13.588578] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:56.257 [2024-12-05 19:00:13.588588] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:56.257 [2024-12-05 19:00:13.590700] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:56.257 [2024-12-05 19:00:13.590734] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:56.257 pt2 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.257 malloc3 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.257 [2024-12-05 19:00:13.617122] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:56.257 [2024-12-05 19:00:13.617232] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:56.257 [2024-12-05 19:00:13.617267] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:56.257 [2024-12-05 19:00:13.617296] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:56.257 [2024-12-05 19:00:13.619409] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:56.257 [2024-12-05 19:00:13.619500] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:56.257 pt3 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:56.257 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.258 malloc4 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.258 [2024-12-05 19:00:13.668765] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:09:56.258 [2024-12-05 19:00:13.668950] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:56.258 [2024-12-05 19:00:13.669028] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:56.258 [2024-12-05 19:00:13.669114] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:56.258 [2024-12-05 19:00:13.673510] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:56.258 [2024-12-05 19:00:13.673613] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:09:56.258 pt4 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.258 [2024-12-05 19:00:13.681901] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:56.258 [2024-12-05 19:00:13.684284] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:56.258 [2024-12-05 19:00:13.684400] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:56.258 [2024-12-05 19:00:13.684499] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:09:56.258 [2024-12-05 19:00:13.684718] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:09:56.258 [2024-12-05 19:00:13.684771] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:56.258 [2024-12-05 19:00:13.685063] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:56.258 [2024-12-05 19:00:13.685251] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:09:56.258 [2024-12-05 19:00:13.685297] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:09:56.258 [2024-12-05 19:00:13.685451] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:56.258 "name": "raid_bdev1", 00:09:56.258 "uuid": "61d77d52-b624-471a-a279-29b1f010bf18", 00:09:56.258 "strip_size_kb": 64, 00:09:56.258 "state": "online", 00:09:56.258 "raid_level": "raid0", 00:09:56.258 "superblock": true, 00:09:56.258 "num_base_bdevs": 4, 00:09:56.258 "num_base_bdevs_discovered": 4, 00:09:56.258 "num_base_bdevs_operational": 4, 00:09:56.258 "base_bdevs_list": [ 00:09:56.258 { 00:09:56.258 "name": "pt1", 00:09:56.258 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:56.258 "is_configured": true, 00:09:56.258 "data_offset": 2048, 00:09:56.258 "data_size": 63488 00:09:56.258 }, 00:09:56.258 { 00:09:56.258 "name": "pt2", 00:09:56.258 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:56.258 "is_configured": true, 00:09:56.258 "data_offset": 2048, 00:09:56.258 "data_size": 63488 00:09:56.258 }, 00:09:56.258 { 00:09:56.258 "name": "pt3", 00:09:56.258 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:56.258 "is_configured": true, 00:09:56.258 "data_offset": 2048, 00:09:56.258 "data_size": 63488 00:09:56.258 }, 00:09:56.258 { 00:09:56.258 "name": "pt4", 00:09:56.258 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:56.258 "is_configured": true, 00:09:56.258 "data_offset": 2048, 00:09:56.258 "data_size": 63488 00:09:56.258 } 00:09:56.258 ] 00:09:56.258 }' 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:56.258 19:00:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:56.830 [2024-12-05 19:00:14.137312] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:56.830 "name": "raid_bdev1", 00:09:56.830 "aliases": [ 00:09:56.830 "61d77d52-b624-471a-a279-29b1f010bf18" 00:09:56.830 ], 00:09:56.830 "product_name": "Raid Volume", 00:09:56.830 "block_size": 512, 00:09:56.830 "num_blocks": 253952, 00:09:56.830 "uuid": "61d77d52-b624-471a-a279-29b1f010bf18", 00:09:56.830 "assigned_rate_limits": { 00:09:56.830 "rw_ios_per_sec": 0, 00:09:56.830 "rw_mbytes_per_sec": 0, 00:09:56.830 "r_mbytes_per_sec": 0, 00:09:56.830 "w_mbytes_per_sec": 0 00:09:56.830 }, 00:09:56.830 "claimed": false, 00:09:56.830 "zoned": false, 00:09:56.830 "supported_io_types": { 00:09:56.830 "read": true, 00:09:56.830 "write": true, 00:09:56.830 "unmap": true, 00:09:56.830 "flush": true, 00:09:56.830 "reset": true, 00:09:56.830 "nvme_admin": false, 00:09:56.830 "nvme_io": false, 00:09:56.830 "nvme_io_md": false, 00:09:56.830 "write_zeroes": true, 00:09:56.830 "zcopy": false, 00:09:56.830 "get_zone_info": false, 00:09:56.830 "zone_management": false, 00:09:56.830 "zone_append": false, 00:09:56.830 "compare": false, 00:09:56.830 "compare_and_write": false, 00:09:56.830 "abort": false, 00:09:56.830 "seek_hole": false, 00:09:56.830 "seek_data": false, 00:09:56.830 "copy": false, 00:09:56.830 "nvme_iov_md": false 00:09:56.830 }, 00:09:56.830 "memory_domains": [ 00:09:56.830 { 00:09:56.830 "dma_device_id": "system", 00:09:56.830 "dma_device_type": 1 00:09:56.830 }, 00:09:56.830 { 00:09:56.830 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:56.830 "dma_device_type": 2 00:09:56.830 }, 00:09:56.830 { 00:09:56.830 "dma_device_id": "system", 00:09:56.830 "dma_device_type": 1 00:09:56.830 }, 00:09:56.830 { 00:09:56.830 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:56.830 "dma_device_type": 2 00:09:56.830 }, 00:09:56.830 { 00:09:56.830 "dma_device_id": "system", 00:09:56.830 "dma_device_type": 1 00:09:56.830 }, 00:09:56.830 { 00:09:56.830 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:56.830 "dma_device_type": 2 00:09:56.830 }, 00:09:56.830 { 00:09:56.830 "dma_device_id": "system", 00:09:56.830 "dma_device_type": 1 00:09:56.830 }, 00:09:56.830 { 00:09:56.830 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:56.830 "dma_device_type": 2 00:09:56.830 } 00:09:56.830 ], 00:09:56.830 "driver_specific": { 00:09:56.830 "raid": { 00:09:56.830 "uuid": "61d77d52-b624-471a-a279-29b1f010bf18", 00:09:56.830 "strip_size_kb": 64, 00:09:56.830 "state": "online", 00:09:56.830 "raid_level": "raid0", 00:09:56.830 "superblock": true, 00:09:56.830 "num_base_bdevs": 4, 00:09:56.830 "num_base_bdevs_discovered": 4, 00:09:56.830 "num_base_bdevs_operational": 4, 00:09:56.830 "base_bdevs_list": [ 00:09:56.830 { 00:09:56.830 "name": "pt1", 00:09:56.830 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:56.830 "is_configured": true, 00:09:56.830 "data_offset": 2048, 00:09:56.830 "data_size": 63488 00:09:56.830 }, 00:09:56.830 { 00:09:56.830 "name": "pt2", 00:09:56.830 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:56.830 "is_configured": true, 00:09:56.830 "data_offset": 2048, 00:09:56.830 "data_size": 63488 00:09:56.830 }, 00:09:56.830 { 00:09:56.830 "name": "pt3", 00:09:56.830 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:56.830 "is_configured": true, 00:09:56.830 "data_offset": 2048, 00:09:56.830 "data_size": 63488 00:09:56.830 }, 00:09:56.830 { 00:09:56.830 "name": "pt4", 00:09:56.830 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:56.830 "is_configured": true, 00:09:56.830 "data_offset": 2048, 00:09:56.830 "data_size": 63488 00:09:56.830 } 00:09:56.830 ] 00:09:56.830 } 00:09:56.830 } 00:09:56.830 }' 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:56.830 pt2 00:09:56.830 pt3 00:09:56.830 pt4' 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:56.830 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:56.831 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.092 [2024-12-05 19:00:14.436757] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=61d77d52-b624-471a-a279-29b1f010bf18 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 61d77d52-b624-471a-a279-29b1f010bf18 ']' 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.092 [2024-12-05 19:00:14.480400] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:57.092 [2024-12-05 19:00:14.480472] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:57.092 [2024-12-05 19:00:14.480560] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:57.092 [2024-12-05 19:00:14.480640] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:57.092 [2024-12-05 19:00:14.480722] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.092 [2024-12-05 19:00:14.636150] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:57.092 [2024-12-05 19:00:14.638085] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:57.092 [2024-12-05 19:00:14.638186] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:57.092 [2024-12-05 19:00:14.638233] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:09:57.092 [2024-12-05 19:00:14.638308] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:57.092 [2024-12-05 19:00:14.638375] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:57.092 [2024-12-05 19:00:14.638419] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:57.092 [2024-12-05 19:00:14.638498] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:09:57.092 [2024-12-05 19:00:14.638545] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:57.092 [2024-12-05 19:00:14.638590] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:09:57.092 request: 00:09:57.092 { 00:09:57.092 "name": "raid_bdev1", 00:09:57.092 "raid_level": "raid0", 00:09:57.092 "base_bdevs": [ 00:09:57.092 "malloc1", 00:09:57.092 "malloc2", 00:09:57.092 "malloc3", 00:09:57.092 "malloc4" 00:09:57.092 ], 00:09:57.092 "strip_size_kb": 64, 00:09:57.092 "superblock": false, 00:09:57.092 "method": "bdev_raid_create", 00:09:57.092 "req_id": 1 00:09:57.092 } 00:09:57.092 Got JSON-RPC error response 00:09:57.092 response: 00:09:57.092 { 00:09:57.092 "code": -17, 00:09:57.092 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:57.092 } 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:09:57.092 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:57.093 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:57.093 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:57.093 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.093 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:57.093 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.093 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.353 [2024-12-05 19:00:14.696028] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:57.353 [2024-12-05 19:00:14.696078] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:57.353 [2024-12-05 19:00:14.696099] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:57.353 [2024-12-05 19:00:14.696107] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:57.353 [2024-12-05 19:00:14.698296] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:57.353 [2024-12-05 19:00:14.698366] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:57.353 [2024-12-05 19:00:14.698435] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:57.353 [2024-12-05 19:00:14.698466] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:57.353 pt1 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:57.353 "name": "raid_bdev1", 00:09:57.353 "uuid": "61d77d52-b624-471a-a279-29b1f010bf18", 00:09:57.353 "strip_size_kb": 64, 00:09:57.353 "state": "configuring", 00:09:57.353 "raid_level": "raid0", 00:09:57.353 "superblock": true, 00:09:57.353 "num_base_bdevs": 4, 00:09:57.353 "num_base_bdevs_discovered": 1, 00:09:57.353 "num_base_bdevs_operational": 4, 00:09:57.353 "base_bdevs_list": [ 00:09:57.353 { 00:09:57.353 "name": "pt1", 00:09:57.353 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:57.353 "is_configured": true, 00:09:57.353 "data_offset": 2048, 00:09:57.353 "data_size": 63488 00:09:57.353 }, 00:09:57.353 { 00:09:57.353 "name": null, 00:09:57.353 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:57.353 "is_configured": false, 00:09:57.353 "data_offset": 2048, 00:09:57.353 "data_size": 63488 00:09:57.353 }, 00:09:57.353 { 00:09:57.353 "name": null, 00:09:57.353 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:57.353 "is_configured": false, 00:09:57.353 "data_offset": 2048, 00:09:57.353 "data_size": 63488 00:09:57.353 }, 00:09:57.353 { 00:09:57.353 "name": null, 00:09:57.353 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:57.353 "is_configured": false, 00:09:57.353 "data_offset": 2048, 00:09:57.353 "data_size": 63488 00:09:57.353 } 00:09:57.353 ] 00:09:57.353 }' 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:57.353 19:00:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.613 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:09:57.613 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:57.613 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.613 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.613 [2024-12-05 19:00:15.083399] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:57.613 [2024-12-05 19:00:15.083490] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:57.613 [2024-12-05 19:00:15.083527] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:09:57.613 [2024-12-05 19:00:15.083553] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:57.613 [2024-12-05 19:00:15.083920] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:57.613 [2024-12-05 19:00:15.083977] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:57.613 [2024-12-05 19:00:15.084057] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:57.613 [2024-12-05 19:00:15.084103] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:57.613 pt2 00:09:57.613 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.613 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.614 [2024-12-05 19:00:15.095403] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:57.614 "name": "raid_bdev1", 00:09:57.614 "uuid": "61d77d52-b624-471a-a279-29b1f010bf18", 00:09:57.614 "strip_size_kb": 64, 00:09:57.614 "state": "configuring", 00:09:57.614 "raid_level": "raid0", 00:09:57.614 "superblock": true, 00:09:57.614 "num_base_bdevs": 4, 00:09:57.614 "num_base_bdevs_discovered": 1, 00:09:57.614 "num_base_bdevs_operational": 4, 00:09:57.614 "base_bdevs_list": [ 00:09:57.614 { 00:09:57.614 "name": "pt1", 00:09:57.614 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:57.614 "is_configured": true, 00:09:57.614 "data_offset": 2048, 00:09:57.614 "data_size": 63488 00:09:57.614 }, 00:09:57.614 { 00:09:57.614 "name": null, 00:09:57.614 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:57.614 "is_configured": false, 00:09:57.614 "data_offset": 0, 00:09:57.614 "data_size": 63488 00:09:57.614 }, 00:09:57.614 { 00:09:57.614 "name": null, 00:09:57.614 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:57.614 "is_configured": false, 00:09:57.614 "data_offset": 2048, 00:09:57.614 "data_size": 63488 00:09:57.614 }, 00:09:57.614 { 00:09:57.614 "name": null, 00:09:57.614 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:57.614 "is_configured": false, 00:09:57.614 "data_offset": 2048, 00:09:57.614 "data_size": 63488 00:09:57.614 } 00:09:57.614 ] 00:09:57.614 }' 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:57.614 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.183 [2024-12-05 19:00:15.518654] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:58.183 [2024-12-05 19:00:15.518721] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:58.183 [2024-12-05 19:00:15.518735] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:09:58.183 [2024-12-05 19:00:15.518745] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:58.183 [2024-12-05 19:00:15.519067] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:58.183 [2024-12-05 19:00:15.519086] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:58.183 [2024-12-05 19:00:15.519138] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:58.183 [2024-12-05 19:00:15.519163] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:58.183 pt2 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.183 [2024-12-05 19:00:15.530610] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:58.183 [2024-12-05 19:00:15.530657] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:58.183 [2024-12-05 19:00:15.530686] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:58.183 [2024-12-05 19:00:15.530696] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:58.183 [2024-12-05 19:00:15.531044] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:58.183 [2024-12-05 19:00:15.531069] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:58.183 [2024-12-05 19:00:15.531117] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:58.183 [2024-12-05 19:00:15.531135] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:58.183 pt3 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.183 [2024-12-05 19:00:15.542599] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:09:58.183 [2024-12-05 19:00:15.542649] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:58.183 [2024-12-05 19:00:15.542661] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:09:58.183 [2024-12-05 19:00:15.542686] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:58.183 [2024-12-05 19:00:15.542975] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:58.183 [2024-12-05 19:00:15.542993] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:09:58.183 [2024-12-05 19:00:15.543042] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:09:58.183 [2024-12-05 19:00:15.543066] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:09:58.183 [2024-12-05 19:00:15.543158] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:58.183 [2024-12-05 19:00:15.543168] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:58.183 [2024-12-05 19:00:15.543389] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:58.183 [2024-12-05 19:00:15.543499] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:58.183 [2024-12-05 19:00:15.543506] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:09:58.183 [2024-12-05 19:00:15.543595] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:58.183 pt4 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:58.183 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:58.184 "name": "raid_bdev1", 00:09:58.184 "uuid": "61d77d52-b624-471a-a279-29b1f010bf18", 00:09:58.184 "strip_size_kb": 64, 00:09:58.184 "state": "online", 00:09:58.184 "raid_level": "raid0", 00:09:58.184 "superblock": true, 00:09:58.184 "num_base_bdevs": 4, 00:09:58.184 "num_base_bdevs_discovered": 4, 00:09:58.184 "num_base_bdevs_operational": 4, 00:09:58.184 "base_bdevs_list": [ 00:09:58.184 { 00:09:58.184 "name": "pt1", 00:09:58.184 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:58.184 "is_configured": true, 00:09:58.184 "data_offset": 2048, 00:09:58.184 "data_size": 63488 00:09:58.184 }, 00:09:58.184 { 00:09:58.184 "name": "pt2", 00:09:58.184 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:58.184 "is_configured": true, 00:09:58.184 "data_offset": 2048, 00:09:58.184 "data_size": 63488 00:09:58.184 }, 00:09:58.184 { 00:09:58.184 "name": "pt3", 00:09:58.184 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:58.184 "is_configured": true, 00:09:58.184 "data_offset": 2048, 00:09:58.184 "data_size": 63488 00:09:58.184 }, 00:09:58.184 { 00:09:58.184 "name": "pt4", 00:09:58.184 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:58.184 "is_configured": true, 00:09:58.184 "data_offset": 2048, 00:09:58.184 "data_size": 63488 00:09:58.184 } 00:09:58.184 ] 00:09:58.184 }' 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:58.184 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.443 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:58.443 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:58.443 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:58.443 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:58.443 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:58.443 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:58.443 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:58.443 19:00:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:58.443 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.443 19:00:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.703 [2024-12-05 19:00:16.006114] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:58.703 "name": "raid_bdev1", 00:09:58.703 "aliases": [ 00:09:58.703 "61d77d52-b624-471a-a279-29b1f010bf18" 00:09:58.703 ], 00:09:58.703 "product_name": "Raid Volume", 00:09:58.703 "block_size": 512, 00:09:58.703 "num_blocks": 253952, 00:09:58.703 "uuid": "61d77d52-b624-471a-a279-29b1f010bf18", 00:09:58.703 "assigned_rate_limits": { 00:09:58.703 "rw_ios_per_sec": 0, 00:09:58.703 "rw_mbytes_per_sec": 0, 00:09:58.703 "r_mbytes_per_sec": 0, 00:09:58.703 "w_mbytes_per_sec": 0 00:09:58.703 }, 00:09:58.703 "claimed": false, 00:09:58.703 "zoned": false, 00:09:58.703 "supported_io_types": { 00:09:58.703 "read": true, 00:09:58.703 "write": true, 00:09:58.703 "unmap": true, 00:09:58.703 "flush": true, 00:09:58.703 "reset": true, 00:09:58.703 "nvme_admin": false, 00:09:58.703 "nvme_io": false, 00:09:58.703 "nvme_io_md": false, 00:09:58.703 "write_zeroes": true, 00:09:58.703 "zcopy": false, 00:09:58.703 "get_zone_info": false, 00:09:58.703 "zone_management": false, 00:09:58.703 "zone_append": false, 00:09:58.703 "compare": false, 00:09:58.703 "compare_and_write": false, 00:09:58.703 "abort": false, 00:09:58.703 "seek_hole": false, 00:09:58.703 "seek_data": false, 00:09:58.703 "copy": false, 00:09:58.703 "nvme_iov_md": false 00:09:58.703 }, 00:09:58.703 "memory_domains": [ 00:09:58.703 { 00:09:58.703 "dma_device_id": "system", 00:09:58.703 "dma_device_type": 1 00:09:58.703 }, 00:09:58.703 { 00:09:58.703 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.703 "dma_device_type": 2 00:09:58.703 }, 00:09:58.703 { 00:09:58.703 "dma_device_id": "system", 00:09:58.703 "dma_device_type": 1 00:09:58.703 }, 00:09:58.703 { 00:09:58.703 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.703 "dma_device_type": 2 00:09:58.703 }, 00:09:58.703 { 00:09:58.703 "dma_device_id": "system", 00:09:58.703 "dma_device_type": 1 00:09:58.703 }, 00:09:58.703 { 00:09:58.703 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.703 "dma_device_type": 2 00:09:58.703 }, 00:09:58.703 { 00:09:58.703 "dma_device_id": "system", 00:09:58.703 "dma_device_type": 1 00:09:58.703 }, 00:09:58.703 { 00:09:58.703 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.703 "dma_device_type": 2 00:09:58.703 } 00:09:58.703 ], 00:09:58.703 "driver_specific": { 00:09:58.703 "raid": { 00:09:58.703 "uuid": "61d77d52-b624-471a-a279-29b1f010bf18", 00:09:58.703 "strip_size_kb": 64, 00:09:58.703 "state": "online", 00:09:58.703 "raid_level": "raid0", 00:09:58.703 "superblock": true, 00:09:58.703 "num_base_bdevs": 4, 00:09:58.703 "num_base_bdevs_discovered": 4, 00:09:58.703 "num_base_bdevs_operational": 4, 00:09:58.703 "base_bdevs_list": [ 00:09:58.703 { 00:09:58.703 "name": "pt1", 00:09:58.703 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:58.703 "is_configured": true, 00:09:58.703 "data_offset": 2048, 00:09:58.703 "data_size": 63488 00:09:58.703 }, 00:09:58.703 { 00:09:58.703 "name": "pt2", 00:09:58.703 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:58.703 "is_configured": true, 00:09:58.703 "data_offset": 2048, 00:09:58.703 "data_size": 63488 00:09:58.703 }, 00:09:58.703 { 00:09:58.703 "name": "pt3", 00:09:58.703 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:58.703 "is_configured": true, 00:09:58.703 "data_offset": 2048, 00:09:58.703 "data_size": 63488 00:09:58.703 }, 00:09:58.703 { 00:09:58.703 "name": "pt4", 00:09:58.703 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:58.703 "is_configured": true, 00:09:58.703 "data_offset": 2048, 00:09:58.703 "data_size": 63488 00:09:58.703 } 00:09:58.703 ] 00:09:58.703 } 00:09:58.703 } 00:09:58.703 }' 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:58.703 pt2 00:09:58.703 pt3 00:09:58.703 pt4' 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:58.703 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.964 [2024-12-05 19:00:16.321533] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 61d77d52-b624-471a-a279-29b1f010bf18 '!=' 61d77d52-b624-471a-a279-29b1f010bf18 ']' 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 80863 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 80863 ']' 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 80863 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 80863 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:58.964 killing process with pid 80863 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 80863' 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 80863 00:09:58.964 [2024-12-05 19:00:16.373819] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:58.964 [2024-12-05 19:00:16.373893] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:58.964 [2024-12-05 19:00:16.373960] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:58.964 [2024-12-05 19:00:16.373972] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:09:58.964 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 80863 00:09:58.964 [2024-12-05 19:00:16.417414] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:59.225 19:00:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:59.225 00:09:59.225 real 0m4.008s 00:09:59.225 user 0m6.331s 00:09:59.225 sys 0m0.857s 00:09:59.225 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:59.225 19:00:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.225 ************************************ 00:09:59.225 END TEST raid_superblock_test 00:09:59.225 ************************************ 00:09:59.225 19:00:16 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 4 read 00:09:59.225 19:00:16 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:09:59.225 19:00:16 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:59.225 19:00:16 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:59.225 ************************************ 00:09:59.225 START TEST raid_read_error_test 00:09:59.225 ************************************ 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 4 read 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.ZZ7uA9AAZ9 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=81086 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 81086 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 81086 ']' 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:59.225 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:59.225 19:00:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.486 [2024-12-05 19:00:16.808361] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:09:59.486 [2024-12-05 19:00:16.808492] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81086 ] 00:09:59.486 [2024-12-05 19:00:16.962533] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:59.486 [2024-12-05 19:00:16.987323] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:59.486 [2024-12-05 19:00:17.030491] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:59.486 [2024-12-05 19:00:17.030533] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.426 BaseBdev1_malloc 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.426 true 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.426 [2024-12-05 19:00:17.650645] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:00.426 [2024-12-05 19:00:17.650726] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:00.426 [2024-12-05 19:00:17.650745] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:00.426 [2024-12-05 19:00:17.650760] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:00.426 [2024-12-05 19:00:17.652909] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:00.426 [2024-12-05 19:00:17.652952] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:00.426 BaseBdev1 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.426 BaseBdev2_malloc 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.426 true 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.426 [2024-12-05 19:00:17.691254] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:00.426 [2024-12-05 19:00:17.691320] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:00.426 [2024-12-05 19:00:17.691337] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:00.426 [2024-12-05 19:00:17.691353] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:00.426 [2024-12-05 19:00:17.693436] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:00.426 [2024-12-05 19:00:17.693471] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:00.426 BaseBdev2 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.426 BaseBdev3_malloc 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.426 true 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.426 [2024-12-05 19:00:17.731911] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:00.426 [2024-12-05 19:00:17.731960] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:00.426 [2024-12-05 19:00:17.731994] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:00.426 [2024-12-05 19:00:17.732003] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:00.426 [2024-12-05 19:00:17.734030] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:00.426 [2024-12-05 19:00:17.734063] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:00.426 BaseBdev3 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.426 BaseBdev4_malloc 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.426 true 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.426 [2024-12-05 19:00:17.781748] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:00.426 [2024-12-05 19:00:17.781792] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:00.426 [2024-12-05 19:00:17.781813] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:00.426 [2024-12-05 19:00:17.781821] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:00.426 [2024-12-05 19:00:17.783861] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:00.426 [2024-12-05 19:00:17.783895] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:00.426 BaseBdev4 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.426 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.426 [2024-12-05 19:00:17.793749] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:00.426 [2024-12-05 19:00:17.795632] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:00.426 [2024-12-05 19:00:17.795722] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:00.426 [2024-12-05 19:00:17.795797] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:00.426 [2024-12-05 19:00:17.795987] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:00.426 [2024-12-05 19:00:17.796008] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:00.426 [2024-12-05 19:00:17.796258] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:10:00.427 [2024-12-05 19:00:17.796400] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:00.427 [2024-12-05 19:00:17.796420] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:00.427 [2024-12-05 19:00:17.796543] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.427 "name": "raid_bdev1", 00:10:00.427 "uuid": "a70c1880-ca1e-46e9-9ad5-aab8210ed06a", 00:10:00.427 "strip_size_kb": 64, 00:10:00.427 "state": "online", 00:10:00.427 "raid_level": "raid0", 00:10:00.427 "superblock": true, 00:10:00.427 "num_base_bdevs": 4, 00:10:00.427 "num_base_bdevs_discovered": 4, 00:10:00.427 "num_base_bdevs_operational": 4, 00:10:00.427 "base_bdevs_list": [ 00:10:00.427 { 00:10:00.427 "name": "BaseBdev1", 00:10:00.427 "uuid": "3197da09-4f07-5e42-a9b1-45c113a973fd", 00:10:00.427 "is_configured": true, 00:10:00.427 "data_offset": 2048, 00:10:00.427 "data_size": 63488 00:10:00.427 }, 00:10:00.427 { 00:10:00.427 "name": "BaseBdev2", 00:10:00.427 "uuid": "999982a5-bcb1-50d2-9394-f3d1f78ed5a7", 00:10:00.427 "is_configured": true, 00:10:00.427 "data_offset": 2048, 00:10:00.427 "data_size": 63488 00:10:00.427 }, 00:10:00.427 { 00:10:00.427 "name": "BaseBdev3", 00:10:00.427 "uuid": "23d66d4e-8cef-5428-98da-aff37da0cdd1", 00:10:00.427 "is_configured": true, 00:10:00.427 "data_offset": 2048, 00:10:00.427 "data_size": 63488 00:10:00.427 }, 00:10:00.427 { 00:10:00.427 "name": "BaseBdev4", 00:10:00.427 "uuid": "6e2da91c-df47-5762-8e28-766a6f48ed5e", 00:10:00.427 "is_configured": true, 00:10:00.427 "data_offset": 2048, 00:10:00.427 "data_size": 63488 00:10:00.427 } 00:10:00.427 ] 00:10:00.427 }' 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.427 19:00:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.686 19:00:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:00.686 19:00:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:00.946 [2024-12-05 19:00:18.329171] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:01.905 "name": "raid_bdev1", 00:10:01.905 "uuid": "a70c1880-ca1e-46e9-9ad5-aab8210ed06a", 00:10:01.905 "strip_size_kb": 64, 00:10:01.905 "state": "online", 00:10:01.905 "raid_level": "raid0", 00:10:01.905 "superblock": true, 00:10:01.905 "num_base_bdevs": 4, 00:10:01.905 "num_base_bdevs_discovered": 4, 00:10:01.905 "num_base_bdevs_operational": 4, 00:10:01.905 "base_bdevs_list": [ 00:10:01.905 { 00:10:01.905 "name": "BaseBdev1", 00:10:01.905 "uuid": "3197da09-4f07-5e42-a9b1-45c113a973fd", 00:10:01.905 "is_configured": true, 00:10:01.905 "data_offset": 2048, 00:10:01.905 "data_size": 63488 00:10:01.905 }, 00:10:01.905 { 00:10:01.905 "name": "BaseBdev2", 00:10:01.905 "uuid": "999982a5-bcb1-50d2-9394-f3d1f78ed5a7", 00:10:01.905 "is_configured": true, 00:10:01.905 "data_offset": 2048, 00:10:01.905 "data_size": 63488 00:10:01.905 }, 00:10:01.905 { 00:10:01.905 "name": "BaseBdev3", 00:10:01.905 "uuid": "23d66d4e-8cef-5428-98da-aff37da0cdd1", 00:10:01.905 "is_configured": true, 00:10:01.905 "data_offset": 2048, 00:10:01.905 "data_size": 63488 00:10:01.905 }, 00:10:01.905 { 00:10:01.905 "name": "BaseBdev4", 00:10:01.905 "uuid": "6e2da91c-df47-5762-8e28-766a6f48ed5e", 00:10:01.905 "is_configured": true, 00:10:01.905 "data_offset": 2048, 00:10:01.905 "data_size": 63488 00:10:01.905 } 00:10:01.905 ] 00:10:01.905 }' 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:01.905 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.166 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:02.166 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:02.166 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.166 [2024-12-05 19:00:19.696931] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:02.166 [2024-12-05 19:00:19.696967] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:02.166 [2024-12-05 19:00:19.699591] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:02.166 [2024-12-05 19:00:19.699644] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:02.166 [2024-12-05 19:00:19.699701] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:02.166 [2024-12-05 19:00:19.699712] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:02.166 { 00:10:02.166 "results": [ 00:10:02.166 { 00:10:02.166 "job": "raid_bdev1", 00:10:02.166 "core_mask": "0x1", 00:10:02.166 "workload": "randrw", 00:10:02.166 "percentage": 50, 00:10:02.166 "status": "finished", 00:10:02.166 "queue_depth": 1, 00:10:02.166 "io_size": 131072, 00:10:02.166 "runtime": 1.368629, 00:10:02.166 "iops": 16854.09267230199, 00:10:02.166 "mibps": 2106.761584037749, 00:10:02.166 "io_failed": 1, 00:10:02.166 "io_timeout": 0, 00:10:02.166 "avg_latency_us": 82.0542114712303, 00:10:02.166 "min_latency_us": 25.041048034934498, 00:10:02.166 "max_latency_us": 1366.5257641921398 00:10:02.166 } 00:10:02.166 ], 00:10:02.166 "core_count": 1 00:10:02.166 } 00:10:02.166 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:02.166 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 81086 00:10:02.166 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 81086 ']' 00:10:02.166 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 81086 00:10:02.166 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:10:02.166 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:02.166 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81086 00:10:02.425 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:02.425 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:02.425 killing process with pid 81086 00:10:02.425 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81086' 00:10:02.425 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 81086 00:10:02.425 [2024-12-05 19:00:19.742740] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:02.425 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 81086 00:10:02.425 [2024-12-05 19:00:19.777986] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:02.685 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:02.685 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.ZZ7uA9AAZ9 00:10:02.685 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:02.685 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:10:02.685 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:10:02.685 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:02.685 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:02.685 19:00:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:10:02.685 00:10:02.685 real 0m3.284s 00:10:02.685 user 0m4.143s 00:10:02.685 sys 0m0.532s 00:10:02.685 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:02.685 19:00:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.685 ************************************ 00:10:02.685 END TEST raid_read_error_test 00:10:02.685 ************************************ 00:10:02.685 19:00:20 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 4 write 00:10:02.685 19:00:20 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:02.685 19:00:20 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:02.685 19:00:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:02.685 ************************************ 00:10:02.685 START TEST raid_write_error_test 00:10:02.685 ************************************ 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid0 4 write 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.EjJVzc6nxy 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=81198 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 81198 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 81198 ']' 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:02.685 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:02.685 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.685 [2024-12-05 19:00:20.156511] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:10:02.685 [2024-12-05 19:00:20.156648] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81198 ] 00:10:02.945 [2024-12-05 19:00:20.309678] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:02.945 [2024-12-05 19:00:20.334322] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:02.945 [2024-12-05 19:00:20.377447] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:02.945 [2024-12-05 19:00:20.377488] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:03.513 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:03.513 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:03.513 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:03.513 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:03.513 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.513 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.513 BaseBdev1_malloc 00:10:03.513 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.513 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:03.513 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.513 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.513 true 00:10:03.513 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.513 19:00:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:03.513 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.513 19:00:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.513 [2024-12-05 19:00:21.005602] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:03.513 [2024-12-05 19:00:21.005653] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:03.513 [2024-12-05 19:00:21.005685] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:03.513 [2024-12-05 19:00:21.005700] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:03.513 [2024-12-05 19:00:21.007903] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:03.513 [2024-12-05 19:00:21.007939] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:03.513 BaseBdev1 00:10:03.513 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.513 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:03.513 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:03.513 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.513 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.513 BaseBdev2_malloc 00:10:03.513 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.513 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:03.513 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.513 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.513 true 00:10:03.513 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.513 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:03.513 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.513 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.513 [2024-12-05 19:00:21.042387] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:03.513 [2024-12-05 19:00:21.042448] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:03.513 [2024-12-05 19:00:21.042465] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:03.513 [2024-12-05 19:00:21.042482] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:03.513 [2024-12-05 19:00:21.044625] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:03.513 [2024-12-05 19:00:21.044672] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:03.513 BaseBdev2 00:10:03.513 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.514 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:03.514 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:03.514 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.514 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.514 BaseBdev3_malloc 00:10:03.514 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.514 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:03.514 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.514 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.514 true 00:10:03.514 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.773 [2024-12-05 19:00:21.075131] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:03.773 [2024-12-05 19:00:21.075191] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:03.773 [2024-12-05 19:00:21.075210] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:03.773 [2024-12-05 19:00:21.075218] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:03.773 [2024-12-05 19:00:21.077229] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:03.773 [2024-12-05 19:00:21.077267] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:03.773 BaseBdev3 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.773 BaseBdev4_malloc 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.773 true 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.773 [2024-12-05 19:00:21.121005] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:03.773 [2024-12-05 19:00:21.121054] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:03.773 [2024-12-05 19:00:21.121090] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:03.773 [2024-12-05 19:00:21.121098] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:03.773 [2024-12-05 19:00:21.123086] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:03.773 [2024-12-05 19:00:21.123119] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:03.773 BaseBdev4 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.773 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.773 [2024-12-05 19:00:21.133027] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:03.773 [2024-12-05 19:00:21.134841] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:03.773 [2024-12-05 19:00:21.134934] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:03.774 [2024-12-05 19:00:21.134984] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:03.774 [2024-12-05 19:00:21.135173] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:03.774 [2024-12-05 19:00:21.135192] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:03.774 [2024-12-05 19:00:21.135450] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:10:03.774 [2024-12-05 19:00:21.135598] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:03.774 [2024-12-05 19:00:21.135618] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:03.774 [2024-12-05 19:00:21.135755] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.774 "name": "raid_bdev1", 00:10:03.774 "uuid": "cf96e2f7-b4cc-4bbf-9a5b-5e2a10296035", 00:10:03.774 "strip_size_kb": 64, 00:10:03.774 "state": "online", 00:10:03.774 "raid_level": "raid0", 00:10:03.774 "superblock": true, 00:10:03.774 "num_base_bdevs": 4, 00:10:03.774 "num_base_bdevs_discovered": 4, 00:10:03.774 "num_base_bdevs_operational": 4, 00:10:03.774 "base_bdevs_list": [ 00:10:03.774 { 00:10:03.774 "name": "BaseBdev1", 00:10:03.774 "uuid": "e5fdf5e3-04ec-584c-b34f-152fb11460df", 00:10:03.774 "is_configured": true, 00:10:03.774 "data_offset": 2048, 00:10:03.774 "data_size": 63488 00:10:03.774 }, 00:10:03.774 { 00:10:03.774 "name": "BaseBdev2", 00:10:03.774 "uuid": "09827195-5e1a-50e5-a45a-0c5fb4935b2e", 00:10:03.774 "is_configured": true, 00:10:03.774 "data_offset": 2048, 00:10:03.774 "data_size": 63488 00:10:03.774 }, 00:10:03.774 { 00:10:03.774 "name": "BaseBdev3", 00:10:03.774 "uuid": "1fc4803e-b3ef-533d-93d3-039db5169b2f", 00:10:03.774 "is_configured": true, 00:10:03.774 "data_offset": 2048, 00:10:03.774 "data_size": 63488 00:10:03.774 }, 00:10:03.774 { 00:10:03.774 "name": "BaseBdev4", 00:10:03.774 "uuid": "7c5c1fd9-8671-5f87-a107-ae691f52181d", 00:10:03.774 "is_configured": true, 00:10:03.774 "data_offset": 2048, 00:10:03.774 "data_size": 63488 00:10:03.774 } 00:10:03.774 ] 00:10:03.774 }' 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.774 19:00:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.032 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:04.032 19:00:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:04.292 [2024-12-05 19:00:21.660463] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.229 19:00:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.230 19:00:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.230 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:05.230 "name": "raid_bdev1", 00:10:05.230 "uuid": "cf96e2f7-b4cc-4bbf-9a5b-5e2a10296035", 00:10:05.230 "strip_size_kb": 64, 00:10:05.230 "state": "online", 00:10:05.230 "raid_level": "raid0", 00:10:05.230 "superblock": true, 00:10:05.230 "num_base_bdevs": 4, 00:10:05.230 "num_base_bdevs_discovered": 4, 00:10:05.230 "num_base_bdevs_operational": 4, 00:10:05.230 "base_bdevs_list": [ 00:10:05.230 { 00:10:05.230 "name": "BaseBdev1", 00:10:05.230 "uuid": "e5fdf5e3-04ec-584c-b34f-152fb11460df", 00:10:05.230 "is_configured": true, 00:10:05.230 "data_offset": 2048, 00:10:05.230 "data_size": 63488 00:10:05.230 }, 00:10:05.230 { 00:10:05.230 "name": "BaseBdev2", 00:10:05.230 "uuid": "09827195-5e1a-50e5-a45a-0c5fb4935b2e", 00:10:05.230 "is_configured": true, 00:10:05.230 "data_offset": 2048, 00:10:05.230 "data_size": 63488 00:10:05.230 }, 00:10:05.230 { 00:10:05.230 "name": "BaseBdev3", 00:10:05.230 "uuid": "1fc4803e-b3ef-533d-93d3-039db5169b2f", 00:10:05.230 "is_configured": true, 00:10:05.230 "data_offset": 2048, 00:10:05.230 "data_size": 63488 00:10:05.230 }, 00:10:05.230 { 00:10:05.230 "name": "BaseBdev4", 00:10:05.230 "uuid": "7c5c1fd9-8671-5f87-a107-ae691f52181d", 00:10:05.230 "is_configured": true, 00:10:05.230 "data_offset": 2048, 00:10:05.230 "data_size": 63488 00:10:05.230 } 00:10:05.230 ] 00:10:05.230 }' 00:10:05.230 19:00:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:05.230 19:00:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.489 19:00:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:05.489 19:00:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:05.489 19:00:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.749 [2024-12-05 19:00:23.052640] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:05.749 [2024-12-05 19:00:23.052686] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:05.749 [2024-12-05 19:00:23.055242] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:05.749 [2024-12-05 19:00:23.055298] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:05.749 [2024-12-05 19:00:23.055347] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:05.749 [2024-12-05 19:00:23.055360] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:05.749 { 00:10:05.749 "results": [ 00:10:05.749 { 00:10:05.749 "job": "raid_bdev1", 00:10:05.750 "core_mask": "0x1", 00:10:05.750 "workload": "randrw", 00:10:05.750 "percentage": 50, 00:10:05.750 "status": "finished", 00:10:05.750 "queue_depth": 1, 00:10:05.750 "io_size": 131072, 00:10:05.750 "runtime": 1.393115, 00:10:05.750 "iops": 16878.72142644362, 00:10:05.750 "mibps": 2109.8401783054524, 00:10:05.750 "io_failed": 1, 00:10:05.750 "io_timeout": 0, 00:10:05.750 "avg_latency_us": 81.96422649484163, 00:10:05.750 "min_latency_us": 25.041048034934498, 00:10:05.750 "max_latency_us": 1366.5257641921398 00:10:05.750 } 00:10:05.750 ], 00:10:05.750 "core_count": 1 00:10:05.750 } 00:10:05.750 19:00:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:05.750 19:00:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 81198 00:10:05.750 19:00:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 81198 ']' 00:10:05.750 19:00:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 81198 00:10:05.750 19:00:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:10:05.750 19:00:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:05.750 19:00:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81198 00:10:05.750 19:00:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:05.750 19:00:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:05.750 killing process with pid 81198 00:10:05.750 19:00:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81198' 00:10:05.750 19:00:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 81198 00:10:05.750 [2024-12-05 19:00:23.090215] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:05.750 19:00:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 81198 00:10:05.750 [2024-12-05 19:00:23.124800] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:06.011 19:00:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.EjJVzc6nxy 00:10:06.011 19:00:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:06.011 19:00:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:06.011 19:00:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:10:06.011 19:00:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:10:06.011 19:00:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:06.011 19:00:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:06.011 19:00:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:10:06.011 00:10:06.011 real 0m3.284s 00:10:06.011 user 0m4.140s 00:10:06.011 sys 0m0.526s 00:10:06.011 19:00:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:06.011 19:00:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.011 ************************************ 00:10:06.011 END TEST raid_write_error_test 00:10:06.011 ************************************ 00:10:06.011 19:00:23 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:10:06.011 19:00:23 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 4 false 00:10:06.011 19:00:23 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:06.011 19:00:23 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:06.011 19:00:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:06.011 ************************************ 00:10:06.011 START TEST raid_state_function_test 00:10:06.011 ************************************ 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 4 false 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=81307 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 81307' 00:10:06.011 Process raid pid: 81307 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 81307 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 81307 ']' 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:06.011 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:06.011 19:00:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.011 [2024-12-05 19:00:23.506783] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:10:06.011 [2024-12-05 19:00:23.506919] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:06.272 [2024-12-05 19:00:23.639073] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:06.272 [2024-12-05 19:00:23.663328] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:06.272 [2024-12-05 19:00:23.706102] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:06.272 [2024-12-05 19:00:23.706139] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.847 [2024-12-05 19:00:24.329218] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:06.847 [2024-12-05 19:00:24.329275] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:06.847 [2024-12-05 19:00:24.329285] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:06.847 [2024-12-05 19:00:24.329295] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:06.847 [2024-12-05 19:00:24.329301] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:06.847 [2024-12-05 19:00:24.329312] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:06.847 [2024-12-05 19:00:24.329318] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:06.847 [2024-12-05 19:00:24.329326] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:06.847 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:06.847 "name": "Existed_Raid", 00:10:06.847 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:06.847 "strip_size_kb": 64, 00:10:06.847 "state": "configuring", 00:10:06.847 "raid_level": "concat", 00:10:06.847 "superblock": false, 00:10:06.847 "num_base_bdevs": 4, 00:10:06.847 "num_base_bdevs_discovered": 0, 00:10:06.847 "num_base_bdevs_operational": 4, 00:10:06.848 "base_bdevs_list": [ 00:10:06.848 { 00:10:06.848 "name": "BaseBdev1", 00:10:06.848 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:06.848 "is_configured": false, 00:10:06.848 "data_offset": 0, 00:10:06.848 "data_size": 0 00:10:06.848 }, 00:10:06.848 { 00:10:06.848 "name": "BaseBdev2", 00:10:06.848 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:06.848 "is_configured": false, 00:10:06.848 "data_offset": 0, 00:10:06.848 "data_size": 0 00:10:06.848 }, 00:10:06.848 { 00:10:06.848 "name": "BaseBdev3", 00:10:06.848 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:06.848 "is_configured": false, 00:10:06.848 "data_offset": 0, 00:10:06.848 "data_size": 0 00:10:06.848 }, 00:10:06.848 { 00:10:06.848 "name": "BaseBdev4", 00:10:06.848 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:06.848 "is_configured": false, 00:10:06.848 "data_offset": 0, 00:10:06.848 "data_size": 0 00:10:06.848 } 00:10:06.848 ] 00:10:06.848 }' 00:10:06.848 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:06.848 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.422 [2024-12-05 19:00:24.724542] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:07.422 [2024-12-05 19:00:24.724583] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.422 [2024-12-05 19:00:24.736558] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:07.422 [2024-12-05 19:00:24.736604] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:07.422 [2024-12-05 19:00:24.736613] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:07.422 [2024-12-05 19:00:24.736621] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:07.422 [2024-12-05 19:00:24.736627] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:07.422 [2024-12-05 19:00:24.736635] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:07.422 [2024-12-05 19:00:24.736640] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:07.422 [2024-12-05 19:00:24.736648] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.422 [2024-12-05 19:00:24.757531] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:07.422 BaseBdev1 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.422 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.422 [ 00:10:07.422 { 00:10:07.422 "name": "BaseBdev1", 00:10:07.422 "aliases": [ 00:10:07.422 "362324e4-f526-4bf0-9803-f4245a38b6c2" 00:10:07.422 ], 00:10:07.422 "product_name": "Malloc disk", 00:10:07.422 "block_size": 512, 00:10:07.422 "num_blocks": 65536, 00:10:07.422 "uuid": "362324e4-f526-4bf0-9803-f4245a38b6c2", 00:10:07.422 "assigned_rate_limits": { 00:10:07.422 "rw_ios_per_sec": 0, 00:10:07.422 "rw_mbytes_per_sec": 0, 00:10:07.422 "r_mbytes_per_sec": 0, 00:10:07.423 "w_mbytes_per_sec": 0 00:10:07.423 }, 00:10:07.423 "claimed": true, 00:10:07.423 "claim_type": "exclusive_write", 00:10:07.423 "zoned": false, 00:10:07.423 "supported_io_types": { 00:10:07.423 "read": true, 00:10:07.423 "write": true, 00:10:07.423 "unmap": true, 00:10:07.423 "flush": true, 00:10:07.423 "reset": true, 00:10:07.423 "nvme_admin": false, 00:10:07.423 "nvme_io": false, 00:10:07.423 "nvme_io_md": false, 00:10:07.423 "write_zeroes": true, 00:10:07.423 "zcopy": true, 00:10:07.423 "get_zone_info": false, 00:10:07.423 "zone_management": false, 00:10:07.423 "zone_append": false, 00:10:07.423 "compare": false, 00:10:07.423 "compare_and_write": false, 00:10:07.423 "abort": true, 00:10:07.423 "seek_hole": false, 00:10:07.423 "seek_data": false, 00:10:07.423 "copy": true, 00:10:07.423 "nvme_iov_md": false 00:10:07.423 }, 00:10:07.423 "memory_domains": [ 00:10:07.423 { 00:10:07.423 "dma_device_id": "system", 00:10:07.423 "dma_device_type": 1 00:10:07.423 }, 00:10:07.423 { 00:10:07.423 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:07.423 "dma_device_type": 2 00:10:07.423 } 00:10:07.423 ], 00:10:07.423 "driver_specific": {} 00:10:07.423 } 00:10:07.423 ] 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:07.423 "name": "Existed_Raid", 00:10:07.423 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:07.423 "strip_size_kb": 64, 00:10:07.423 "state": "configuring", 00:10:07.423 "raid_level": "concat", 00:10:07.423 "superblock": false, 00:10:07.423 "num_base_bdevs": 4, 00:10:07.423 "num_base_bdevs_discovered": 1, 00:10:07.423 "num_base_bdevs_operational": 4, 00:10:07.423 "base_bdevs_list": [ 00:10:07.423 { 00:10:07.423 "name": "BaseBdev1", 00:10:07.423 "uuid": "362324e4-f526-4bf0-9803-f4245a38b6c2", 00:10:07.423 "is_configured": true, 00:10:07.423 "data_offset": 0, 00:10:07.423 "data_size": 65536 00:10:07.423 }, 00:10:07.423 { 00:10:07.423 "name": "BaseBdev2", 00:10:07.423 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:07.423 "is_configured": false, 00:10:07.423 "data_offset": 0, 00:10:07.423 "data_size": 0 00:10:07.423 }, 00:10:07.423 { 00:10:07.423 "name": "BaseBdev3", 00:10:07.423 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:07.423 "is_configured": false, 00:10:07.423 "data_offset": 0, 00:10:07.423 "data_size": 0 00:10:07.423 }, 00:10:07.423 { 00:10:07.423 "name": "BaseBdev4", 00:10:07.423 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:07.423 "is_configured": false, 00:10:07.423 "data_offset": 0, 00:10:07.423 "data_size": 0 00:10:07.423 } 00:10:07.423 ] 00:10:07.423 }' 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:07.423 19:00:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.691 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:07.691 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.691 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.691 [2024-12-05 19:00:25.240735] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:07.691 [2024-12-05 19:00:25.240780] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:07.691 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.691 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:07.691 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.691 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.952 [2024-12-05 19:00:25.252758] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:07.952 [2024-12-05 19:00:25.254564] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:07.952 [2024-12-05 19:00:25.254609] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:07.952 [2024-12-05 19:00:25.254619] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:07.952 [2024-12-05 19:00:25.254627] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:07.952 [2024-12-05 19:00:25.254633] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:07.952 [2024-12-05 19:00:25.254641] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:07.952 "name": "Existed_Raid", 00:10:07.952 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:07.952 "strip_size_kb": 64, 00:10:07.952 "state": "configuring", 00:10:07.952 "raid_level": "concat", 00:10:07.952 "superblock": false, 00:10:07.952 "num_base_bdevs": 4, 00:10:07.952 "num_base_bdevs_discovered": 1, 00:10:07.952 "num_base_bdevs_operational": 4, 00:10:07.952 "base_bdevs_list": [ 00:10:07.952 { 00:10:07.952 "name": "BaseBdev1", 00:10:07.952 "uuid": "362324e4-f526-4bf0-9803-f4245a38b6c2", 00:10:07.952 "is_configured": true, 00:10:07.952 "data_offset": 0, 00:10:07.952 "data_size": 65536 00:10:07.952 }, 00:10:07.952 { 00:10:07.952 "name": "BaseBdev2", 00:10:07.952 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:07.952 "is_configured": false, 00:10:07.952 "data_offset": 0, 00:10:07.952 "data_size": 0 00:10:07.952 }, 00:10:07.952 { 00:10:07.952 "name": "BaseBdev3", 00:10:07.952 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:07.952 "is_configured": false, 00:10:07.952 "data_offset": 0, 00:10:07.952 "data_size": 0 00:10:07.952 }, 00:10:07.952 { 00:10:07.952 "name": "BaseBdev4", 00:10:07.952 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:07.952 "is_configured": false, 00:10:07.952 "data_offset": 0, 00:10:07.952 "data_size": 0 00:10:07.952 } 00:10:07.952 ] 00:10:07.952 }' 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:07.952 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.211 [2024-12-05 19:00:25.667017] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:08.211 BaseBdev2 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.211 [ 00:10:08.211 { 00:10:08.211 "name": "BaseBdev2", 00:10:08.211 "aliases": [ 00:10:08.211 "26bc2a33-7443-476a-b3b5-71e36e6207b6" 00:10:08.211 ], 00:10:08.211 "product_name": "Malloc disk", 00:10:08.211 "block_size": 512, 00:10:08.211 "num_blocks": 65536, 00:10:08.211 "uuid": "26bc2a33-7443-476a-b3b5-71e36e6207b6", 00:10:08.211 "assigned_rate_limits": { 00:10:08.211 "rw_ios_per_sec": 0, 00:10:08.211 "rw_mbytes_per_sec": 0, 00:10:08.211 "r_mbytes_per_sec": 0, 00:10:08.211 "w_mbytes_per_sec": 0 00:10:08.211 }, 00:10:08.211 "claimed": true, 00:10:08.211 "claim_type": "exclusive_write", 00:10:08.211 "zoned": false, 00:10:08.211 "supported_io_types": { 00:10:08.211 "read": true, 00:10:08.211 "write": true, 00:10:08.211 "unmap": true, 00:10:08.211 "flush": true, 00:10:08.211 "reset": true, 00:10:08.211 "nvme_admin": false, 00:10:08.211 "nvme_io": false, 00:10:08.211 "nvme_io_md": false, 00:10:08.211 "write_zeroes": true, 00:10:08.211 "zcopy": true, 00:10:08.211 "get_zone_info": false, 00:10:08.211 "zone_management": false, 00:10:08.211 "zone_append": false, 00:10:08.211 "compare": false, 00:10:08.211 "compare_and_write": false, 00:10:08.211 "abort": true, 00:10:08.211 "seek_hole": false, 00:10:08.211 "seek_data": false, 00:10:08.211 "copy": true, 00:10:08.211 "nvme_iov_md": false 00:10:08.211 }, 00:10:08.211 "memory_domains": [ 00:10:08.211 { 00:10:08.211 "dma_device_id": "system", 00:10:08.211 "dma_device_type": 1 00:10:08.211 }, 00:10:08.211 { 00:10:08.211 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:08.211 "dma_device_type": 2 00:10:08.211 } 00:10:08.211 ], 00:10:08.211 "driver_specific": {} 00:10:08.211 } 00:10:08.211 ] 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:08.211 "name": "Existed_Raid", 00:10:08.211 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:08.211 "strip_size_kb": 64, 00:10:08.211 "state": "configuring", 00:10:08.211 "raid_level": "concat", 00:10:08.211 "superblock": false, 00:10:08.211 "num_base_bdevs": 4, 00:10:08.211 "num_base_bdevs_discovered": 2, 00:10:08.211 "num_base_bdevs_operational": 4, 00:10:08.211 "base_bdevs_list": [ 00:10:08.211 { 00:10:08.211 "name": "BaseBdev1", 00:10:08.211 "uuid": "362324e4-f526-4bf0-9803-f4245a38b6c2", 00:10:08.211 "is_configured": true, 00:10:08.211 "data_offset": 0, 00:10:08.211 "data_size": 65536 00:10:08.211 }, 00:10:08.211 { 00:10:08.211 "name": "BaseBdev2", 00:10:08.211 "uuid": "26bc2a33-7443-476a-b3b5-71e36e6207b6", 00:10:08.211 "is_configured": true, 00:10:08.211 "data_offset": 0, 00:10:08.211 "data_size": 65536 00:10:08.211 }, 00:10:08.211 { 00:10:08.211 "name": "BaseBdev3", 00:10:08.211 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:08.211 "is_configured": false, 00:10:08.211 "data_offset": 0, 00:10:08.211 "data_size": 0 00:10:08.211 }, 00:10:08.211 { 00:10:08.211 "name": "BaseBdev4", 00:10:08.211 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:08.211 "is_configured": false, 00:10:08.211 "data_offset": 0, 00:10:08.211 "data_size": 0 00:10:08.211 } 00:10:08.211 ] 00:10:08.211 }' 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:08.211 19:00:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.829 [2024-12-05 19:00:26.138178] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:08.829 BaseBdev3 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.829 [ 00:10:08.829 { 00:10:08.829 "name": "BaseBdev3", 00:10:08.829 "aliases": [ 00:10:08.829 "b2951ae8-c030-4e36-b95b-02bd7450737f" 00:10:08.829 ], 00:10:08.829 "product_name": "Malloc disk", 00:10:08.829 "block_size": 512, 00:10:08.829 "num_blocks": 65536, 00:10:08.829 "uuid": "b2951ae8-c030-4e36-b95b-02bd7450737f", 00:10:08.829 "assigned_rate_limits": { 00:10:08.829 "rw_ios_per_sec": 0, 00:10:08.829 "rw_mbytes_per_sec": 0, 00:10:08.829 "r_mbytes_per_sec": 0, 00:10:08.829 "w_mbytes_per_sec": 0 00:10:08.829 }, 00:10:08.829 "claimed": true, 00:10:08.829 "claim_type": "exclusive_write", 00:10:08.829 "zoned": false, 00:10:08.829 "supported_io_types": { 00:10:08.829 "read": true, 00:10:08.829 "write": true, 00:10:08.829 "unmap": true, 00:10:08.829 "flush": true, 00:10:08.829 "reset": true, 00:10:08.829 "nvme_admin": false, 00:10:08.829 "nvme_io": false, 00:10:08.829 "nvme_io_md": false, 00:10:08.829 "write_zeroes": true, 00:10:08.829 "zcopy": true, 00:10:08.829 "get_zone_info": false, 00:10:08.829 "zone_management": false, 00:10:08.829 "zone_append": false, 00:10:08.829 "compare": false, 00:10:08.829 "compare_and_write": false, 00:10:08.829 "abort": true, 00:10:08.829 "seek_hole": false, 00:10:08.829 "seek_data": false, 00:10:08.829 "copy": true, 00:10:08.829 "nvme_iov_md": false 00:10:08.829 }, 00:10:08.829 "memory_domains": [ 00:10:08.829 { 00:10:08.829 "dma_device_id": "system", 00:10:08.829 "dma_device_type": 1 00:10:08.829 }, 00:10:08.829 { 00:10:08.829 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:08.829 "dma_device_type": 2 00:10:08.829 } 00:10:08.829 ], 00:10:08.829 "driver_specific": {} 00:10:08.829 } 00:10:08.829 ] 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:08.829 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:08.829 "name": "Existed_Raid", 00:10:08.829 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:08.829 "strip_size_kb": 64, 00:10:08.829 "state": "configuring", 00:10:08.829 "raid_level": "concat", 00:10:08.829 "superblock": false, 00:10:08.829 "num_base_bdevs": 4, 00:10:08.829 "num_base_bdevs_discovered": 3, 00:10:08.829 "num_base_bdevs_operational": 4, 00:10:08.829 "base_bdevs_list": [ 00:10:08.829 { 00:10:08.829 "name": "BaseBdev1", 00:10:08.829 "uuid": "362324e4-f526-4bf0-9803-f4245a38b6c2", 00:10:08.829 "is_configured": true, 00:10:08.829 "data_offset": 0, 00:10:08.829 "data_size": 65536 00:10:08.829 }, 00:10:08.829 { 00:10:08.829 "name": "BaseBdev2", 00:10:08.830 "uuid": "26bc2a33-7443-476a-b3b5-71e36e6207b6", 00:10:08.830 "is_configured": true, 00:10:08.830 "data_offset": 0, 00:10:08.830 "data_size": 65536 00:10:08.830 }, 00:10:08.830 { 00:10:08.830 "name": "BaseBdev3", 00:10:08.830 "uuid": "b2951ae8-c030-4e36-b95b-02bd7450737f", 00:10:08.830 "is_configured": true, 00:10:08.830 "data_offset": 0, 00:10:08.830 "data_size": 65536 00:10:08.830 }, 00:10:08.830 { 00:10:08.830 "name": "BaseBdev4", 00:10:08.830 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:08.830 "is_configured": false, 00:10:08.830 "data_offset": 0, 00:10:08.830 "data_size": 0 00:10:08.830 } 00:10:08.830 ] 00:10:08.830 }' 00:10:08.830 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:08.830 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.089 [2024-12-05 19:00:26.620312] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:09.089 [2024-12-05 19:00:26.620359] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:09.089 [2024-12-05 19:00:26.620367] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:09.089 [2024-12-05 19:00:26.620650] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:09.089 [2024-12-05 19:00:26.620802] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:09.089 [2024-12-05 19:00:26.620836] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:09.089 [2024-12-05 19:00:26.621060] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:09.089 BaseBdev4 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.089 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.089 [ 00:10:09.089 { 00:10:09.089 "name": "BaseBdev4", 00:10:09.089 "aliases": [ 00:10:09.089 "f87e580b-b74e-4260-b15a-44246572b1db" 00:10:09.089 ], 00:10:09.089 "product_name": "Malloc disk", 00:10:09.089 "block_size": 512, 00:10:09.089 "num_blocks": 65536, 00:10:09.348 "uuid": "f87e580b-b74e-4260-b15a-44246572b1db", 00:10:09.348 "assigned_rate_limits": { 00:10:09.348 "rw_ios_per_sec": 0, 00:10:09.348 "rw_mbytes_per_sec": 0, 00:10:09.348 "r_mbytes_per_sec": 0, 00:10:09.348 "w_mbytes_per_sec": 0 00:10:09.348 }, 00:10:09.348 "claimed": true, 00:10:09.348 "claim_type": "exclusive_write", 00:10:09.348 "zoned": false, 00:10:09.348 "supported_io_types": { 00:10:09.348 "read": true, 00:10:09.348 "write": true, 00:10:09.348 "unmap": true, 00:10:09.348 "flush": true, 00:10:09.348 "reset": true, 00:10:09.348 "nvme_admin": false, 00:10:09.348 "nvme_io": false, 00:10:09.348 "nvme_io_md": false, 00:10:09.348 "write_zeroes": true, 00:10:09.348 "zcopy": true, 00:10:09.348 "get_zone_info": false, 00:10:09.348 "zone_management": false, 00:10:09.348 "zone_append": false, 00:10:09.348 "compare": false, 00:10:09.348 "compare_and_write": false, 00:10:09.348 "abort": true, 00:10:09.348 "seek_hole": false, 00:10:09.348 "seek_data": false, 00:10:09.348 "copy": true, 00:10:09.348 "nvme_iov_md": false 00:10:09.348 }, 00:10:09.348 "memory_domains": [ 00:10:09.348 { 00:10:09.348 "dma_device_id": "system", 00:10:09.348 "dma_device_type": 1 00:10:09.348 }, 00:10:09.348 { 00:10:09.348 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.348 "dma_device_type": 2 00:10:09.348 } 00:10:09.348 ], 00:10:09.348 "driver_specific": {} 00:10:09.348 } 00:10:09.348 ] 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.348 "name": "Existed_Raid", 00:10:09.348 "uuid": "3c9176eb-daf5-4be0-bdf6-4998eee509a8", 00:10:09.348 "strip_size_kb": 64, 00:10:09.348 "state": "online", 00:10:09.348 "raid_level": "concat", 00:10:09.348 "superblock": false, 00:10:09.348 "num_base_bdevs": 4, 00:10:09.348 "num_base_bdevs_discovered": 4, 00:10:09.348 "num_base_bdevs_operational": 4, 00:10:09.348 "base_bdevs_list": [ 00:10:09.348 { 00:10:09.348 "name": "BaseBdev1", 00:10:09.348 "uuid": "362324e4-f526-4bf0-9803-f4245a38b6c2", 00:10:09.348 "is_configured": true, 00:10:09.348 "data_offset": 0, 00:10:09.348 "data_size": 65536 00:10:09.348 }, 00:10:09.348 { 00:10:09.348 "name": "BaseBdev2", 00:10:09.348 "uuid": "26bc2a33-7443-476a-b3b5-71e36e6207b6", 00:10:09.348 "is_configured": true, 00:10:09.348 "data_offset": 0, 00:10:09.348 "data_size": 65536 00:10:09.348 }, 00:10:09.348 { 00:10:09.348 "name": "BaseBdev3", 00:10:09.348 "uuid": "b2951ae8-c030-4e36-b95b-02bd7450737f", 00:10:09.348 "is_configured": true, 00:10:09.348 "data_offset": 0, 00:10:09.348 "data_size": 65536 00:10:09.348 }, 00:10:09.348 { 00:10:09.348 "name": "BaseBdev4", 00:10:09.348 "uuid": "f87e580b-b74e-4260-b15a-44246572b1db", 00:10:09.348 "is_configured": true, 00:10:09.348 "data_offset": 0, 00:10:09.348 "data_size": 65536 00:10:09.348 } 00:10:09.348 ] 00:10:09.348 }' 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.348 19:00:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.607 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:09.607 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:09.607 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:09.607 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:09.607 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:09.607 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:09.607 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:09.607 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:09.607 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.607 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.607 [2024-12-05 19:00:27.083867] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:09.607 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.607 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:09.607 "name": "Existed_Raid", 00:10:09.607 "aliases": [ 00:10:09.607 "3c9176eb-daf5-4be0-bdf6-4998eee509a8" 00:10:09.607 ], 00:10:09.607 "product_name": "Raid Volume", 00:10:09.607 "block_size": 512, 00:10:09.607 "num_blocks": 262144, 00:10:09.607 "uuid": "3c9176eb-daf5-4be0-bdf6-4998eee509a8", 00:10:09.607 "assigned_rate_limits": { 00:10:09.607 "rw_ios_per_sec": 0, 00:10:09.607 "rw_mbytes_per_sec": 0, 00:10:09.607 "r_mbytes_per_sec": 0, 00:10:09.607 "w_mbytes_per_sec": 0 00:10:09.607 }, 00:10:09.607 "claimed": false, 00:10:09.607 "zoned": false, 00:10:09.607 "supported_io_types": { 00:10:09.607 "read": true, 00:10:09.607 "write": true, 00:10:09.607 "unmap": true, 00:10:09.607 "flush": true, 00:10:09.607 "reset": true, 00:10:09.607 "nvme_admin": false, 00:10:09.607 "nvme_io": false, 00:10:09.607 "nvme_io_md": false, 00:10:09.607 "write_zeroes": true, 00:10:09.607 "zcopy": false, 00:10:09.607 "get_zone_info": false, 00:10:09.607 "zone_management": false, 00:10:09.607 "zone_append": false, 00:10:09.607 "compare": false, 00:10:09.607 "compare_and_write": false, 00:10:09.607 "abort": false, 00:10:09.607 "seek_hole": false, 00:10:09.607 "seek_data": false, 00:10:09.607 "copy": false, 00:10:09.607 "nvme_iov_md": false 00:10:09.607 }, 00:10:09.607 "memory_domains": [ 00:10:09.607 { 00:10:09.607 "dma_device_id": "system", 00:10:09.607 "dma_device_type": 1 00:10:09.607 }, 00:10:09.607 { 00:10:09.607 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.607 "dma_device_type": 2 00:10:09.607 }, 00:10:09.607 { 00:10:09.607 "dma_device_id": "system", 00:10:09.607 "dma_device_type": 1 00:10:09.607 }, 00:10:09.607 { 00:10:09.607 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.607 "dma_device_type": 2 00:10:09.607 }, 00:10:09.607 { 00:10:09.607 "dma_device_id": "system", 00:10:09.607 "dma_device_type": 1 00:10:09.607 }, 00:10:09.607 { 00:10:09.607 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.607 "dma_device_type": 2 00:10:09.607 }, 00:10:09.607 { 00:10:09.607 "dma_device_id": "system", 00:10:09.607 "dma_device_type": 1 00:10:09.607 }, 00:10:09.607 { 00:10:09.607 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.607 "dma_device_type": 2 00:10:09.607 } 00:10:09.607 ], 00:10:09.607 "driver_specific": { 00:10:09.607 "raid": { 00:10:09.607 "uuid": "3c9176eb-daf5-4be0-bdf6-4998eee509a8", 00:10:09.607 "strip_size_kb": 64, 00:10:09.607 "state": "online", 00:10:09.607 "raid_level": "concat", 00:10:09.607 "superblock": false, 00:10:09.607 "num_base_bdevs": 4, 00:10:09.607 "num_base_bdevs_discovered": 4, 00:10:09.607 "num_base_bdevs_operational": 4, 00:10:09.607 "base_bdevs_list": [ 00:10:09.607 { 00:10:09.607 "name": "BaseBdev1", 00:10:09.607 "uuid": "362324e4-f526-4bf0-9803-f4245a38b6c2", 00:10:09.607 "is_configured": true, 00:10:09.607 "data_offset": 0, 00:10:09.607 "data_size": 65536 00:10:09.607 }, 00:10:09.607 { 00:10:09.607 "name": "BaseBdev2", 00:10:09.607 "uuid": "26bc2a33-7443-476a-b3b5-71e36e6207b6", 00:10:09.607 "is_configured": true, 00:10:09.607 "data_offset": 0, 00:10:09.607 "data_size": 65536 00:10:09.607 }, 00:10:09.607 { 00:10:09.607 "name": "BaseBdev3", 00:10:09.607 "uuid": "b2951ae8-c030-4e36-b95b-02bd7450737f", 00:10:09.607 "is_configured": true, 00:10:09.607 "data_offset": 0, 00:10:09.607 "data_size": 65536 00:10:09.607 }, 00:10:09.608 { 00:10:09.608 "name": "BaseBdev4", 00:10:09.608 "uuid": "f87e580b-b74e-4260-b15a-44246572b1db", 00:10:09.608 "is_configured": true, 00:10:09.608 "data_offset": 0, 00:10:09.608 "data_size": 65536 00:10:09.608 } 00:10:09.608 ] 00:10:09.608 } 00:10:09.608 } 00:10:09.608 }' 00:10:09.608 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:09.608 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:09.608 BaseBdev2 00:10:09.608 BaseBdev3 00:10:09.608 BaseBdev4' 00:10:09.608 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.866 [2024-12-05 19:00:27.311192] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:09.866 [2024-12-05 19:00:27.311221] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:09.866 [2024-12-05 19:00:27.311267] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:09.866 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.867 "name": "Existed_Raid", 00:10:09.867 "uuid": "3c9176eb-daf5-4be0-bdf6-4998eee509a8", 00:10:09.867 "strip_size_kb": 64, 00:10:09.867 "state": "offline", 00:10:09.867 "raid_level": "concat", 00:10:09.867 "superblock": false, 00:10:09.867 "num_base_bdevs": 4, 00:10:09.867 "num_base_bdevs_discovered": 3, 00:10:09.867 "num_base_bdevs_operational": 3, 00:10:09.867 "base_bdevs_list": [ 00:10:09.867 { 00:10:09.867 "name": null, 00:10:09.867 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.867 "is_configured": false, 00:10:09.867 "data_offset": 0, 00:10:09.867 "data_size": 65536 00:10:09.867 }, 00:10:09.867 { 00:10:09.867 "name": "BaseBdev2", 00:10:09.867 "uuid": "26bc2a33-7443-476a-b3b5-71e36e6207b6", 00:10:09.867 "is_configured": true, 00:10:09.867 "data_offset": 0, 00:10:09.867 "data_size": 65536 00:10:09.867 }, 00:10:09.867 { 00:10:09.867 "name": "BaseBdev3", 00:10:09.867 "uuid": "b2951ae8-c030-4e36-b95b-02bd7450737f", 00:10:09.867 "is_configured": true, 00:10:09.867 "data_offset": 0, 00:10:09.867 "data_size": 65536 00:10:09.867 }, 00:10:09.867 { 00:10:09.867 "name": "BaseBdev4", 00:10:09.867 "uuid": "f87e580b-b74e-4260-b15a-44246572b1db", 00:10:09.867 "is_configured": true, 00:10:09.867 "data_offset": 0, 00:10:09.867 "data_size": 65536 00:10:09.867 } 00:10:09.867 ] 00:10:09.867 }' 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.867 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.436 [2024-12-05 19:00:27.829624] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.436 [2024-12-05 19:00:27.896637] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.436 [2024-12-05 19:00:27.951418] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:10.436 [2024-12-05 19:00:27.951508] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.436 19:00:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.697 19:00:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.697 BaseBdev2 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.697 [ 00:10:10.697 { 00:10:10.697 "name": "BaseBdev2", 00:10:10.697 "aliases": [ 00:10:10.697 "83df4879-a197-4e58-9115-bd64263acd30" 00:10:10.697 ], 00:10:10.697 "product_name": "Malloc disk", 00:10:10.697 "block_size": 512, 00:10:10.697 "num_blocks": 65536, 00:10:10.697 "uuid": "83df4879-a197-4e58-9115-bd64263acd30", 00:10:10.697 "assigned_rate_limits": { 00:10:10.697 "rw_ios_per_sec": 0, 00:10:10.697 "rw_mbytes_per_sec": 0, 00:10:10.697 "r_mbytes_per_sec": 0, 00:10:10.697 "w_mbytes_per_sec": 0 00:10:10.697 }, 00:10:10.697 "claimed": false, 00:10:10.697 "zoned": false, 00:10:10.697 "supported_io_types": { 00:10:10.697 "read": true, 00:10:10.697 "write": true, 00:10:10.697 "unmap": true, 00:10:10.697 "flush": true, 00:10:10.697 "reset": true, 00:10:10.697 "nvme_admin": false, 00:10:10.697 "nvme_io": false, 00:10:10.697 "nvme_io_md": false, 00:10:10.697 "write_zeroes": true, 00:10:10.697 "zcopy": true, 00:10:10.697 "get_zone_info": false, 00:10:10.697 "zone_management": false, 00:10:10.697 "zone_append": false, 00:10:10.697 "compare": false, 00:10:10.697 "compare_and_write": false, 00:10:10.697 "abort": true, 00:10:10.697 "seek_hole": false, 00:10:10.697 "seek_data": false, 00:10:10.697 "copy": true, 00:10:10.697 "nvme_iov_md": false 00:10:10.697 }, 00:10:10.697 "memory_domains": [ 00:10:10.697 { 00:10:10.697 "dma_device_id": "system", 00:10:10.697 "dma_device_type": 1 00:10:10.697 }, 00:10:10.697 { 00:10:10.697 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:10.697 "dma_device_type": 2 00:10:10.697 } 00:10:10.697 ], 00:10:10.697 "driver_specific": {} 00:10:10.697 } 00:10:10.697 ] 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.697 BaseBdev3 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:10.697 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.698 [ 00:10:10.698 { 00:10:10.698 "name": "BaseBdev3", 00:10:10.698 "aliases": [ 00:10:10.698 "938a722a-d389-47b1-b47c-be287a95dab8" 00:10:10.698 ], 00:10:10.698 "product_name": "Malloc disk", 00:10:10.698 "block_size": 512, 00:10:10.698 "num_blocks": 65536, 00:10:10.698 "uuid": "938a722a-d389-47b1-b47c-be287a95dab8", 00:10:10.698 "assigned_rate_limits": { 00:10:10.698 "rw_ios_per_sec": 0, 00:10:10.698 "rw_mbytes_per_sec": 0, 00:10:10.698 "r_mbytes_per_sec": 0, 00:10:10.698 "w_mbytes_per_sec": 0 00:10:10.698 }, 00:10:10.698 "claimed": false, 00:10:10.698 "zoned": false, 00:10:10.698 "supported_io_types": { 00:10:10.698 "read": true, 00:10:10.698 "write": true, 00:10:10.698 "unmap": true, 00:10:10.698 "flush": true, 00:10:10.698 "reset": true, 00:10:10.698 "nvme_admin": false, 00:10:10.698 "nvme_io": false, 00:10:10.698 "nvme_io_md": false, 00:10:10.698 "write_zeroes": true, 00:10:10.698 "zcopy": true, 00:10:10.698 "get_zone_info": false, 00:10:10.698 "zone_management": false, 00:10:10.698 "zone_append": false, 00:10:10.698 "compare": false, 00:10:10.698 "compare_and_write": false, 00:10:10.698 "abort": true, 00:10:10.698 "seek_hole": false, 00:10:10.698 "seek_data": false, 00:10:10.698 "copy": true, 00:10:10.698 "nvme_iov_md": false 00:10:10.698 }, 00:10:10.698 "memory_domains": [ 00:10:10.698 { 00:10:10.698 "dma_device_id": "system", 00:10:10.698 "dma_device_type": 1 00:10:10.698 }, 00:10:10.698 { 00:10:10.698 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:10.698 "dma_device_type": 2 00:10:10.698 } 00:10:10.698 ], 00:10:10.698 "driver_specific": {} 00:10:10.698 } 00:10:10.698 ] 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.698 BaseBdev4 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.698 [ 00:10:10.698 { 00:10:10.698 "name": "BaseBdev4", 00:10:10.698 "aliases": [ 00:10:10.698 "0f3e2ff2-1233-49c0-9138-25c8162d84f8" 00:10:10.698 ], 00:10:10.698 "product_name": "Malloc disk", 00:10:10.698 "block_size": 512, 00:10:10.698 "num_blocks": 65536, 00:10:10.698 "uuid": "0f3e2ff2-1233-49c0-9138-25c8162d84f8", 00:10:10.698 "assigned_rate_limits": { 00:10:10.698 "rw_ios_per_sec": 0, 00:10:10.698 "rw_mbytes_per_sec": 0, 00:10:10.698 "r_mbytes_per_sec": 0, 00:10:10.698 "w_mbytes_per_sec": 0 00:10:10.698 }, 00:10:10.698 "claimed": false, 00:10:10.698 "zoned": false, 00:10:10.698 "supported_io_types": { 00:10:10.698 "read": true, 00:10:10.698 "write": true, 00:10:10.698 "unmap": true, 00:10:10.698 "flush": true, 00:10:10.698 "reset": true, 00:10:10.698 "nvme_admin": false, 00:10:10.698 "nvme_io": false, 00:10:10.698 "nvme_io_md": false, 00:10:10.698 "write_zeroes": true, 00:10:10.698 "zcopy": true, 00:10:10.698 "get_zone_info": false, 00:10:10.698 "zone_management": false, 00:10:10.698 "zone_append": false, 00:10:10.698 "compare": false, 00:10:10.698 "compare_and_write": false, 00:10:10.698 "abort": true, 00:10:10.698 "seek_hole": false, 00:10:10.698 "seek_data": false, 00:10:10.698 "copy": true, 00:10:10.698 "nvme_iov_md": false 00:10:10.698 }, 00:10:10.698 "memory_domains": [ 00:10:10.698 { 00:10:10.698 "dma_device_id": "system", 00:10:10.698 "dma_device_type": 1 00:10:10.698 }, 00:10:10.698 { 00:10:10.698 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:10.698 "dma_device_type": 2 00:10:10.698 } 00:10:10.698 ], 00:10:10.698 "driver_specific": {} 00:10:10.698 } 00:10:10.698 ] 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.698 [2024-12-05 19:00:28.166798] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:10.698 [2024-12-05 19:00:28.166898] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:10.698 [2024-12-05 19:00:28.166958] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:10.698 [2024-12-05 19:00:28.168760] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:10.698 [2024-12-05 19:00:28.168847] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:10.698 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:10.698 "name": "Existed_Raid", 00:10:10.698 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:10.698 "strip_size_kb": 64, 00:10:10.698 "state": "configuring", 00:10:10.698 "raid_level": "concat", 00:10:10.698 "superblock": false, 00:10:10.698 "num_base_bdevs": 4, 00:10:10.698 "num_base_bdevs_discovered": 3, 00:10:10.698 "num_base_bdevs_operational": 4, 00:10:10.698 "base_bdevs_list": [ 00:10:10.698 { 00:10:10.698 "name": "BaseBdev1", 00:10:10.698 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:10.698 "is_configured": false, 00:10:10.698 "data_offset": 0, 00:10:10.698 "data_size": 0 00:10:10.698 }, 00:10:10.698 { 00:10:10.698 "name": "BaseBdev2", 00:10:10.698 "uuid": "83df4879-a197-4e58-9115-bd64263acd30", 00:10:10.698 "is_configured": true, 00:10:10.698 "data_offset": 0, 00:10:10.698 "data_size": 65536 00:10:10.699 }, 00:10:10.699 { 00:10:10.699 "name": "BaseBdev3", 00:10:10.699 "uuid": "938a722a-d389-47b1-b47c-be287a95dab8", 00:10:10.699 "is_configured": true, 00:10:10.699 "data_offset": 0, 00:10:10.699 "data_size": 65536 00:10:10.699 }, 00:10:10.699 { 00:10:10.699 "name": "BaseBdev4", 00:10:10.699 "uuid": "0f3e2ff2-1233-49c0-9138-25c8162d84f8", 00:10:10.699 "is_configured": true, 00:10:10.699 "data_offset": 0, 00:10:10.699 "data_size": 65536 00:10:10.699 } 00:10:10.699 ] 00:10:10.699 }' 00:10:10.699 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:10.699 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.269 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:11.269 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.269 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.269 [2024-12-05 19:00:28.594055] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:11.269 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.269 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:11.269 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:11.269 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:11.269 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:11.269 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:11.269 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:11.269 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.269 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.269 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.270 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.270 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.270 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:11.270 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.270 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.270 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.270 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.270 "name": "Existed_Raid", 00:10:11.270 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:11.270 "strip_size_kb": 64, 00:10:11.270 "state": "configuring", 00:10:11.270 "raid_level": "concat", 00:10:11.270 "superblock": false, 00:10:11.270 "num_base_bdevs": 4, 00:10:11.270 "num_base_bdevs_discovered": 2, 00:10:11.270 "num_base_bdevs_operational": 4, 00:10:11.270 "base_bdevs_list": [ 00:10:11.270 { 00:10:11.270 "name": "BaseBdev1", 00:10:11.270 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:11.270 "is_configured": false, 00:10:11.270 "data_offset": 0, 00:10:11.270 "data_size": 0 00:10:11.270 }, 00:10:11.270 { 00:10:11.270 "name": null, 00:10:11.270 "uuid": "83df4879-a197-4e58-9115-bd64263acd30", 00:10:11.270 "is_configured": false, 00:10:11.270 "data_offset": 0, 00:10:11.270 "data_size": 65536 00:10:11.270 }, 00:10:11.270 { 00:10:11.270 "name": "BaseBdev3", 00:10:11.270 "uuid": "938a722a-d389-47b1-b47c-be287a95dab8", 00:10:11.270 "is_configured": true, 00:10:11.270 "data_offset": 0, 00:10:11.270 "data_size": 65536 00:10:11.270 }, 00:10:11.270 { 00:10:11.270 "name": "BaseBdev4", 00:10:11.270 "uuid": "0f3e2ff2-1233-49c0-9138-25c8162d84f8", 00:10:11.270 "is_configured": true, 00:10:11.270 "data_offset": 0, 00:10:11.270 "data_size": 65536 00:10:11.270 } 00:10:11.270 ] 00:10:11.270 }' 00:10:11.270 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.270 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.530 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.530 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:11.530 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.530 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.531 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.531 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:11.531 19:00:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:11.531 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.531 19:00:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.531 [2024-12-05 19:00:29.004351] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:11.531 BaseBdev1 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.531 [ 00:10:11.531 { 00:10:11.531 "name": "BaseBdev1", 00:10:11.531 "aliases": [ 00:10:11.531 "001b4524-0f2e-419e-b9eb-6e0a78ee5bf5" 00:10:11.531 ], 00:10:11.531 "product_name": "Malloc disk", 00:10:11.531 "block_size": 512, 00:10:11.531 "num_blocks": 65536, 00:10:11.531 "uuid": "001b4524-0f2e-419e-b9eb-6e0a78ee5bf5", 00:10:11.531 "assigned_rate_limits": { 00:10:11.531 "rw_ios_per_sec": 0, 00:10:11.531 "rw_mbytes_per_sec": 0, 00:10:11.531 "r_mbytes_per_sec": 0, 00:10:11.531 "w_mbytes_per_sec": 0 00:10:11.531 }, 00:10:11.531 "claimed": true, 00:10:11.531 "claim_type": "exclusive_write", 00:10:11.531 "zoned": false, 00:10:11.531 "supported_io_types": { 00:10:11.531 "read": true, 00:10:11.531 "write": true, 00:10:11.531 "unmap": true, 00:10:11.531 "flush": true, 00:10:11.531 "reset": true, 00:10:11.531 "nvme_admin": false, 00:10:11.531 "nvme_io": false, 00:10:11.531 "nvme_io_md": false, 00:10:11.531 "write_zeroes": true, 00:10:11.531 "zcopy": true, 00:10:11.531 "get_zone_info": false, 00:10:11.531 "zone_management": false, 00:10:11.531 "zone_append": false, 00:10:11.531 "compare": false, 00:10:11.531 "compare_and_write": false, 00:10:11.531 "abort": true, 00:10:11.531 "seek_hole": false, 00:10:11.531 "seek_data": false, 00:10:11.531 "copy": true, 00:10:11.531 "nvme_iov_md": false 00:10:11.531 }, 00:10:11.531 "memory_domains": [ 00:10:11.531 { 00:10:11.531 "dma_device_id": "system", 00:10:11.531 "dma_device_type": 1 00:10:11.531 }, 00:10:11.531 { 00:10:11.531 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.531 "dma_device_type": 2 00:10:11.531 } 00:10:11.531 ], 00:10:11.531 "driver_specific": {} 00:10:11.531 } 00:10:11.531 ] 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.531 "name": "Existed_Raid", 00:10:11.531 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:11.531 "strip_size_kb": 64, 00:10:11.531 "state": "configuring", 00:10:11.531 "raid_level": "concat", 00:10:11.531 "superblock": false, 00:10:11.531 "num_base_bdevs": 4, 00:10:11.531 "num_base_bdevs_discovered": 3, 00:10:11.531 "num_base_bdevs_operational": 4, 00:10:11.531 "base_bdevs_list": [ 00:10:11.531 { 00:10:11.531 "name": "BaseBdev1", 00:10:11.531 "uuid": "001b4524-0f2e-419e-b9eb-6e0a78ee5bf5", 00:10:11.531 "is_configured": true, 00:10:11.531 "data_offset": 0, 00:10:11.531 "data_size": 65536 00:10:11.531 }, 00:10:11.531 { 00:10:11.531 "name": null, 00:10:11.531 "uuid": "83df4879-a197-4e58-9115-bd64263acd30", 00:10:11.531 "is_configured": false, 00:10:11.531 "data_offset": 0, 00:10:11.531 "data_size": 65536 00:10:11.531 }, 00:10:11.531 { 00:10:11.531 "name": "BaseBdev3", 00:10:11.531 "uuid": "938a722a-d389-47b1-b47c-be287a95dab8", 00:10:11.531 "is_configured": true, 00:10:11.531 "data_offset": 0, 00:10:11.531 "data_size": 65536 00:10:11.531 }, 00:10:11.531 { 00:10:11.531 "name": "BaseBdev4", 00:10:11.531 "uuid": "0f3e2ff2-1233-49c0-9138-25c8162d84f8", 00:10:11.531 "is_configured": true, 00:10:11.531 "data_offset": 0, 00:10:11.531 "data_size": 65536 00:10:11.531 } 00:10:11.531 ] 00:10:11.531 }' 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.531 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.103 [2024-12-05 19:00:29.451640] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:12.103 "name": "Existed_Raid", 00:10:12.103 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:12.103 "strip_size_kb": 64, 00:10:12.103 "state": "configuring", 00:10:12.103 "raid_level": "concat", 00:10:12.103 "superblock": false, 00:10:12.103 "num_base_bdevs": 4, 00:10:12.103 "num_base_bdevs_discovered": 2, 00:10:12.103 "num_base_bdevs_operational": 4, 00:10:12.103 "base_bdevs_list": [ 00:10:12.103 { 00:10:12.103 "name": "BaseBdev1", 00:10:12.103 "uuid": "001b4524-0f2e-419e-b9eb-6e0a78ee5bf5", 00:10:12.103 "is_configured": true, 00:10:12.103 "data_offset": 0, 00:10:12.103 "data_size": 65536 00:10:12.103 }, 00:10:12.103 { 00:10:12.103 "name": null, 00:10:12.103 "uuid": "83df4879-a197-4e58-9115-bd64263acd30", 00:10:12.103 "is_configured": false, 00:10:12.103 "data_offset": 0, 00:10:12.103 "data_size": 65536 00:10:12.103 }, 00:10:12.103 { 00:10:12.103 "name": null, 00:10:12.103 "uuid": "938a722a-d389-47b1-b47c-be287a95dab8", 00:10:12.103 "is_configured": false, 00:10:12.103 "data_offset": 0, 00:10:12.103 "data_size": 65536 00:10:12.103 }, 00:10:12.103 { 00:10:12.103 "name": "BaseBdev4", 00:10:12.103 "uuid": "0f3e2ff2-1233-49c0-9138-25c8162d84f8", 00:10:12.103 "is_configured": true, 00:10:12.103 "data_offset": 0, 00:10:12.103 "data_size": 65536 00:10:12.103 } 00:10:12.103 ] 00:10:12.103 }' 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:12.103 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.365 [2024-12-05 19:00:29.783075] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:12.365 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:12.366 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:12.366 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:12.366 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:12.366 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:12.366 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.366 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.366 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.366 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:12.366 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.366 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:12.366 "name": "Existed_Raid", 00:10:12.366 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:12.366 "strip_size_kb": 64, 00:10:12.366 "state": "configuring", 00:10:12.366 "raid_level": "concat", 00:10:12.366 "superblock": false, 00:10:12.366 "num_base_bdevs": 4, 00:10:12.366 "num_base_bdevs_discovered": 3, 00:10:12.366 "num_base_bdevs_operational": 4, 00:10:12.366 "base_bdevs_list": [ 00:10:12.366 { 00:10:12.366 "name": "BaseBdev1", 00:10:12.366 "uuid": "001b4524-0f2e-419e-b9eb-6e0a78ee5bf5", 00:10:12.366 "is_configured": true, 00:10:12.366 "data_offset": 0, 00:10:12.366 "data_size": 65536 00:10:12.366 }, 00:10:12.366 { 00:10:12.366 "name": null, 00:10:12.366 "uuid": "83df4879-a197-4e58-9115-bd64263acd30", 00:10:12.366 "is_configured": false, 00:10:12.366 "data_offset": 0, 00:10:12.366 "data_size": 65536 00:10:12.366 }, 00:10:12.366 { 00:10:12.366 "name": "BaseBdev3", 00:10:12.366 "uuid": "938a722a-d389-47b1-b47c-be287a95dab8", 00:10:12.366 "is_configured": true, 00:10:12.366 "data_offset": 0, 00:10:12.366 "data_size": 65536 00:10:12.366 }, 00:10:12.366 { 00:10:12.366 "name": "BaseBdev4", 00:10:12.366 "uuid": "0f3e2ff2-1233-49c0-9138-25c8162d84f8", 00:10:12.366 "is_configured": true, 00:10:12.366 "data_offset": 0, 00:10:12.366 "data_size": 65536 00:10:12.366 } 00:10:12.366 ] 00:10:12.366 }' 00:10:12.366 19:00:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:12.366 19:00:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.626 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:12.626 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.626 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.626 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.626 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.886 [2024-12-05 19:00:30.202431] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:12.886 "name": "Existed_Raid", 00:10:12.886 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:12.886 "strip_size_kb": 64, 00:10:12.886 "state": "configuring", 00:10:12.886 "raid_level": "concat", 00:10:12.886 "superblock": false, 00:10:12.886 "num_base_bdevs": 4, 00:10:12.886 "num_base_bdevs_discovered": 2, 00:10:12.886 "num_base_bdevs_operational": 4, 00:10:12.886 "base_bdevs_list": [ 00:10:12.886 { 00:10:12.886 "name": null, 00:10:12.886 "uuid": "001b4524-0f2e-419e-b9eb-6e0a78ee5bf5", 00:10:12.886 "is_configured": false, 00:10:12.886 "data_offset": 0, 00:10:12.886 "data_size": 65536 00:10:12.886 }, 00:10:12.886 { 00:10:12.886 "name": null, 00:10:12.886 "uuid": "83df4879-a197-4e58-9115-bd64263acd30", 00:10:12.886 "is_configured": false, 00:10:12.886 "data_offset": 0, 00:10:12.886 "data_size": 65536 00:10:12.886 }, 00:10:12.886 { 00:10:12.886 "name": "BaseBdev3", 00:10:12.886 "uuid": "938a722a-d389-47b1-b47c-be287a95dab8", 00:10:12.886 "is_configured": true, 00:10:12.886 "data_offset": 0, 00:10:12.886 "data_size": 65536 00:10:12.886 }, 00:10:12.886 { 00:10:12.886 "name": "BaseBdev4", 00:10:12.886 "uuid": "0f3e2ff2-1233-49c0-9138-25c8162d84f8", 00:10:12.886 "is_configured": true, 00:10:12.886 "data_offset": 0, 00:10:12.886 "data_size": 65536 00:10:12.886 } 00:10:12.886 ] 00:10:12.886 }' 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:12.886 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.147 [2024-12-05 19:00:30.640221] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:13.147 "name": "Existed_Raid", 00:10:13.147 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:13.147 "strip_size_kb": 64, 00:10:13.147 "state": "configuring", 00:10:13.147 "raid_level": "concat", 00:10:13.147 "superblock": false, 00:10:13.147 "num_base_bdevs": 4, 00:10:13.147 "num_base_bdevs_discovered": 3, 00:10:13.147 "num_base_bdevs_operational": 4, 00:10:13.147 "base_bdevs_list": [ 00:10:13.147 { 00:10:13.147 "name": null, 00:10:13.147 "uuid": "001b4524-0f2e-419e-b9eb-6e0a78ee5bf5", 00:10:13.147 "is_configured": false, 00:10:13.147 "data_offset": 0, 00:10:13.147 "data_size": 65536 00:10:13.147 }, 00:10:13.147 { 00:10:13.147 "name": "BaseBdev2", 00:10:13.147 "uuid": "83df4879-a197-4e58-9115-bd64263acd30", 00:10:13.147 "is_configured": true, 00:10:13.147 "data_offset": 0, 00:10:13.147 "data_size": 65536 00:10:13.147 }, 00:10:13.147 { 00:10:13.147 "name": "BaseBdev3", 00:10:13.147 "uuid": "938a722a-d389-47b1-b47c-be287a95dab8", 00:10:13.147 "is_configured": true, 00:10:13.147 "data_offset": 0, 00:10:13.147 "data_size": 65536 00:10:13.147 }, 00:10:13.147 { 00:10:13.147 "name": "BaseBdev4", 00:10:13.147 "uuid": "0f3e2ff2-1233-49c0-9138-25c8162d84f8", 00:10:13.147 "is_configured": true, 00:10:13.147 "data_offset": 0, 00:10:13.147 "data_size": 65536 00:10:13.147 } 00:10:13.147 ] 00:10:13.147 }' 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:13.147 19:00:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 001b4524-0f2e-419e-b9eb-6e0a78ee5bf5 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.718 [2024-12-05 19:00:31.138394] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:13.718 [2024-12-05 19:00:31.138436] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:13.718 [2024-12-05 19:00:31.138444] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:13.718 [2024-12-05 19:00:31.138713] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:13.718 [2024-12-05 19:00:31.138848] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:13.718 [2024-12-05 19:00:31.138859] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:13.718 [2024-12-05 19:00:31.139023] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:13.718 NewBaseBdev 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.718 [ 00:10:13.718 { 00:10:13.718 "name": "NewBaseBdev", 00:10:13.718 "aliases": [ 00:10:13.718 "001b4524-0f2e-419e-b9eb-6e0a78ee5bf5" 00:10:13.718 ], 00:10:13.718 "product_name": "Malloc disk", 00:10:13.718 "block_size": 512, 00:10:13.718 "num_blocks": 65536, 00:10:13.718 "uuid": "001b4524-0f2e-419e-b9eb-6e0a78ee5bf5", 00:10:13.718 "assigned_rate_limits": { 00:10:13.718 "rw_ios_per_sec": 0, 00:10:13.718 "rw_mbytes_per_sec": 0, 00:10:13.718 "r_mbytes_per_sec": 0, 00:10:13.718 "w_mbytes_per_sec": 0 00:10:13.718 }, 00:10:13.718 "claimed": true, 00:10:13.718 "claim_type": "exclusive_write", 00:10:13.718 "zoned": false, 00:10:13.718 "supported_io_types": { 00:10:13.718 "read": true, 00:10:13.718 "write": true, 00:10:13.718 "unmap": true, 00:10:13.718 "flush": true, 00:10:13.718 "reset": true, 00:10:13.718 "nvme_admin": false, 00:10:13.718 "nvme_io": false, 00:10:13.718 "nvme_io_md": false, 00:10:13.718 "write_zeroes": true, 00:10:13.718 "zcopy": true, 00:10:13.718 "get_zone_info": false, 00:10:13.718 "zone_management": false, 00:10:13.718 "zone_append": false, 00:10:13.718 "compare": false, 00:10:13.718 "compare_and_write": false, 00:10:13.718 "abort": true, 00:10:13.718 "seek_hole": false, 00:10:13.718 "seek_data": false, 00:10:13.718 "copy": true, 00:10:13.718 "nvme_iov_md": false 00:10:13.718 }, 00:10:13.718 "memory_domains": [ 00:10:13.718 { 00:10:13.718 "dma_device_id": "system", 00:10:13.718 "dma_device_type": 1 00:10:13.718 }, 00:10:13.718 { 00:10:13.718 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:13.718 "dma_device_type": 2 00:10:13.718 } 00:10:13.718 ], 00:10:13.718 "driver_specific": {} 00:10:13.718 } 00:10:13.718 ] 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:13.718 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:13.719 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:13.719 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:13.719 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:13.719 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.719 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.719 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.719 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:13.719 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.719 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:13.719 "name": "Existed_Raid", 00:10:13.719 "uuid": "012e1911-c309-4456-98a7-78ca441cd37c", 00:10:13.719 "strip_size_kb": 64, 00:10:13.719 "state": "online", 00:10:13.719 "raid_level": "concat", 00:10:13.719 "superblock": false, 00:10:13.719 "num_base_bdevs": 4, 00:10:13.719 "num_base_bdevs_discovered": 4, 00:10:13.719 "num_base_bdevs_operational": 4, 00:10:13.719 "base_bdevs_list": [ 00:10:13.719 { 00:10:13.719 "name": "NewBaseBdev", 00:10:13.719 "uuid": "001b4524-0f2e-419e-b9eb-6e0a78ee5bf5", 00:10:13.719 "is_configured": true, 00:10:13.719 "data_offset": 0, 00:10:13.719 "data_size": 65536 00:10:13.719 }, 00:10:13.719 { 00:10:13.719 "name": "BaseBdev2", 00:10:13.719 "uuid": "83df4879-a197-4e58-9115-bd64263acd30", 00:10:13.719 "is_configured": true, 00:10:13.719 "data_offset": 0, 00:10:13.719 "data_size": 65536 00:10:13.719 }, 00:10:13.719 { 00:10:13.719 "name": "BaseBdev3", 00:10:13.719 "uuid": "938a722a-d389-47b1-b47c-be287a95dab8", 00:10:13.719 "is_configured": true, 00:10:13.719 "data_offset": 0, 00:10:13.719 "data_size": 65536 00:10:13.719 }, 00:10:13.719 { 00:10:13.719 "name": "BaseBdev4", 00:10:13.719 "uuid": "0f3e2ff2-1233-49c0-9138-25c8162d84f8", 00:10:13.719 "is_configured": true, 00:10:13.719 "data_offset": 0, 00:10:13.719 "data_size": 65536 00:10:13.719 } 00:10:13.719 ] 00:10:13.719 }' 00:10:13.719 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:13.719 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:14.290 [2024-12-05 19:00:31.550043] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:14.290 "name": "Existed_Raid", 00:10:14.290 "aliases": [ 00:10:14.290 "012e1911-c309-4456-98a7-78ca441cd37c" 00:10:14.290 ], 00:10:14.290 "product_name": "Raid Volume", 00:10:14.290 "block_size": 512, 00:10:14.290 "num_blocks": 262144, 00:10:14.290 "uuid": "012e1911-c309-4456-98a7-78ca441cd37c", 00:10:14.290 "assigned_rate_limits": { 00:10:14.290 "rw_ios_per_sec": 0, 00:10:14.290 "rw_mbytes_per_sec": 0, 00:10:14.290 "r_mbytes_per_sec": 0, 00:10:14.290 "w_mbytes_per_sec": 0 00:10:14.290 }, 00:10:14.290 "claimed": false, 00:10:14.290 "zoned": false, 00:10:14.290 "supported_io_types": { 00:10:14.290 "read": true, 00:10:14.290 "write": true, 00:10:14.290 "unmap": true, 00:10:14.290 "flush": true, 00:10:14.290 "reset": true, 00:10:14.290 "nvme_admin": false, 00:10:14.290 "nvme_io": false, 00:10:14.290 "nvme_io_md": false, 00:10:14.290 "write_zeroes": true, 00:10:14.290 "zcopy": false, 00:10:14.290 "get_zone_info": false, 00:10:14.290 "zone_management": false, 00:10:14.290 "zone_append": false, 00:10:14.290 "compare": false, 00:10:14.290 "compare_and_write": false, 00:10:14.290 "abort": false, 00:10:14.290 "seek_hole": false, 00:10:14.290 "seek_data": false, 00:10:14.290 "copy": false, 00:10:14.290 "nvme_iov_md": false 00:10:14.290 }, 00:10:14.290 "memory_domains": [ 00:10:14.290 { 00:10:14.290 "dma_device_id": "system", 00:10:14.290 "dma_device_type": 1 00:10:14.290 }, 00:10:14.290 { 00:10:14.290 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:14.290 "dma_device_type": 2 00:10:14.290 }, 00:10:14.290 { 00:10:14.290 "dma_device_id": "system", 00:10:14.290 "dma_device_type": 1 00:10:14.290 }, 00:10:14.290 { 00:10:14.290 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:14.290 "dma_device_type": 2 00:10:14.290 }, 00:10:14.290 { 00:10:14.290 "dma_device_id": "system", 00:10:14.290 "dma_device_type": 1 00:10:14.290 }, 00:10:14.290 { 00:10:14.290 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:14.290 "dma_device_type": 2 00:10:14.290 }, 00:10:14.290 { 00:10:14.290 "dma_device_id": "system", 00:10:14.290 "dma_device_type": 1 00:10:14.290 }, 00:10:14.290 { 00:10:14.290 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:14.290 "dma_device_type": 2 00:10:14.290 } 00:10:14.290 ], 00:10:14.290 "driver_specific": { 00:10:14.290 "raid": { 00:10:14.290 "uuid": "012e1911-c309-4456-98a7-78ca441cd37c", 00:10:14.290 "strip_size_kb": 64, 00:10:14.290 "state": "online", 00:10:14.290 "raid_level": "concat", 00:10:14.290 "superblock": false, 00:10:14.290 "num_base_bdevs": 4, 00:10:14.290 "num_base_bdevs_discovered": 4, 00:10:14.290 "num_base_bdevs_operational": 4, 00:10:14.290 "base_bdevs_list": [ 00:10:14.290 { 00:10:14.290 "name": "NewBaseBdev", 00:10:14.290 "uuid": "001b4524-0f2e-419e-b9eb-6e0a78ee5bf5", 00:10:14.290 "is_configured": true, 00:10:14.290 "data_offset": 0, 00:10:14.290 "data_size": 65536 00:10:14.290 }, 00:10:14.290 { 00:10:14.290 "name": "BaseBdev2", 00:10:14.290 "uuid": "83df4879-a197-4e58-9115-bd64263acd30", 00:10:14.290 "is_configured": true, 00:10:14.290 "data_offset": 0, 00:10:14.290 "data_size": 65536 00:10:14.290 }, 00:10:14.290 { 00:10:14.290 "name": "BaseBdev3", 00:10:14.290 "uuid": "938a722a-d389-47b1-b47c-be287a95dab8", 00:10:14.290 "is_configured": true, 00:10:14.290 "data_offset": 0, 00:10:14.290 "data_size": 65536 00:10:14.290 }, 00:10:14.290 { 00:10:14.290 "name": "BaseBdev4", 00:10:14.290 "uuid": "0f3e2ff2-1233-49c0-9138-25c8162d84f8", 00:10:14.290 "is_configured": true, 00:10:14.290 "data_offset": 0, 00:10:14.290 "data_size": 65536 00:10:14.290 } 00:10:14.290 ] 00:10:14.290 } 00:10:14.290 } 00:10:14.290 }' 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:14.290 BaseBdev2 00:10:14.290 BaseBdev3 00:10:14.290 BaseBdev4' 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:14.290 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.291 [2024-12-05 19:00:31.841207] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:14.291 [2024-12-05 19:00:31.841277] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:14.291 [2024-12-05 19:00:31.841384] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:14.291 [2024-12-05 19:00:31.841477] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:14.291 [2024-12-05 19:00:31.841526] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 81307 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 81307 ']' 00:10:14.291 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 81307 00:10:14.551 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:10:14.551 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:14.551 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81307 00:10:14.551 killing process with pid 81307 00:10:14.551 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:14.551 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:14.551 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81307' 00:10:14.551 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 81307 00:10:14.551 [2024-12-05 19:00:31.869404] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:14.551 19:00:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 81307 00:10:14.551 [2024-12-05 19:00:31.909469] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:14.812 00:10:14.812 real 0m8.710s 00:10:14.812 user 0m14.858s 00:10:14.812 sys 0m1.718s 00:10:14.812 ************************************ 00:10:14.812 END TEST raid_state_function_test 00:10:14.812 ************************************ 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.812 19:00:32 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 4 true 00:10:14.812 19:00:32 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:14.812 19:00:32 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:14.812 19:00:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:14.812 ************************************ 00:10:14.812 START TEST raid_state_function_test_sb 00:10:14.812 ************************************ 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test concat 4 true 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=81897 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 81897' 00:10:14.812 Process raid pid: 81897 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 81897 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 81897 ']' 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:14.812 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:14.812 19:00:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.812 [2024-12-05 19:00:32.291072] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:10:14.813 [2024-12-05 19:00:32.291288] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:15.072 [2024-12-05 19:00:32.423640] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:15.072 [2024-12-05 19:00:32.448443] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:15.072 [2024-12-05 19:00:32.491872] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:15.072 [2024-12-05 19:00:32.491989] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.644 [2024-12-05 19:00:33.115206] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:15.644 [2024-12-05 19:00:33.115328] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:15.644 [2024-12-05 19:00:33.115376] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:15.644 [2024-12-05 19:00:33.115401] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:15.644 [2024-12-05 19:00:33.115418] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:15.644 [2024-12-05 19:00:33.115441] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:15.644 [2024-12-05 19:00:33.115458] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:15.644 [2024-12-05 19:00:33.115477] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:15.644 "name": "Existed_Raid", 00:10:15.644 "uuid": "23f46b6a-90c3-461a-94a5-88f4f55ee3bb", 00:10:15.644 "strip_size_kb": 64, 00:10:15.644 "state": "configuring", 00:10:15.644 "raid_level": "concat", 00:10:15.644 "superblock": true, 00:10:15.644 "num_base_bdevs": 4, 00:10:15.644 "num_base_bdevs_discovered": 0, 00:10:15.644 "num_base_bdevs_operational": 4, 00:10:15.644 "base_bdevs_list": [ 00:10:15.644 { 00:10:15.644 "name": "BaseBdev1", 00:10:15.644 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.644 "is_configured": false, 00:10:15.644 "data_offset": 0, 00:10:15.644 "data_size": 0 00:10:15.644 }, 00:10:15.644 { 00:10:15.644 "name": "BaseBdev2", 00:10:15.644 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.644 "is_configured": false, 00:10:15.644 "data_offset": 0, 00:10:15.644 "data_size": 0 00:10:15.644 }, 00:10:15.644 { 00:10:15.644 "name": "BaseBdev3", 00:10:15.644 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.644 "is_configured": false, 00:10:15.644 "data_offset": 0, 00:10:15.644 "data_size": 0 00:10:15.644 }, 00:10:15.644 { 00:10:15.644 "name": "BaseBdev4", 00:10:15.644 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:15.644 "is_configured": false, 00:10:15.644 "data_offset": 0, 00:10:15.644 "data_size": 0 00:10:15.644 } 00:10:15.644 ] 00:10:15.644 }' 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:15.644 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.923 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:15.923 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.923 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.923 [2024-12-05 19:00:33.462509] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:15.923 [2024-12-05 19:00:33.462546] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:15.923 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.923 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:15.923 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.923 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.923 [2024-12-05 19:00:33.474517] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:15.923 [2024-12-05 19:00:33.474557] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:15.923 [2024-12-05 19:00:33.474566] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:15.923 [2024-12-05 19:00:33.474574] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:15.923 [2024-12-05 19:00:33.474580] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:15.923 [2024-12-05 19:00:33.474588] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:15.923 [2024-12-05 19:00:33.474593] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:15.923 [2024-12-05 19:00:33.474602] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:15.923 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.923 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:15.923 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.923 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.182 [2024-12-05 19:00:33.495485] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:16.182 BaseBdev1 00:10:16.182 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.182 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:16.182 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:16.182 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:16.182 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:16.182 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:16.182 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:16.182 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:16.182 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.182 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.182 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.182 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:16.182 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.182 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.182 [ 00:10:16.182 { 00:10:16.182 "name": "BaseBdev1", 00:10:16.182 "aliases": [ 00:10:16.182 "7003e355-7be5-49fa-8fb6-e6e7972f63fa" 00:10:16.182 ], 00:10:16.182 "product_name": "Malloc disk", 00:10:16.182 "block_size": 512, 00:10:16.182 "num_blocks": 65536, 00:10:16.182 "uuid": "7003e355-7be5-49fa-8fb6-e6e7972f63fa", 00:10:16.182 "assigned_rate_limits": { 00:10:16.182 "rw_ios_per_sec": 0, 00:10:16.182 "rw_mbytes_per_sec": 0, 00:10:16.182 "r_mbytes_per_sec": 0, 00:10:16.182 "w_mbytes_per_sec": 0 00:10:16.182 }, 00:10:16.182 "claimed": true, 00:10:16.182 "claim_type": "exclusive_write", 00:10:16.183 "zoned": false, 00:10:16.183 "supported_io_types": { 00:10:16.183 "read": true, 00:10:16.183 "write": true, 00:10:16.183 "unmap": true, 00:10:16.183 "flush": true, 00:10:16.183 "reset": true, 00:10:16.183 "nvme_admin": false, 00:10:16.183 "nvme_io": false, 00:10:16.183 "nvme_io_md": false, 00:10:16.183 "write_zeroes": true, 00:10:16.183 "zcopy": true, 00:10:16.183 "get_zone_info": false, 00:10:16.183 "zone_management": false, 00:10:16.183 "zone_append": false, 00:10:16.183 "compare": false, 00:10:16.183 "compare_and_write": false, 00:10:16.183 "abort": true, 00:10:16.183 "seek_hole": false, 00:10:16.183 "seek_data": false, 00:10:16.183 "copy": true, 00:10:16.183 "nvme_iov_md": false 00:10:16.183 }, 00:10:16.183 "memory_domains": [ 00:10:16.183 { 00:10:16.183 "dma_device_id": "system", 00:10:16.183 "dma_device_type": 1 00:10:16.183 }, 00:10:16.183 { 00:10:16.183 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:16.183 "dma_device_type": 2 00:10:16.183 } 00:10:16.183 ], 00:10:16.183 "driver_specific": {} 00:10:16.183 } 00:10:16.183 ] 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:16.183 "name": "Existed_Raid", 00:10:16.183 "uuid": "a2e86aac-f9c7-4684-b725-73b06221ca83", 00:10:16.183 "strip_size_kb": 64, 00:10:16.183 "state": "configuring", 00:10:16.183 "raid_level": "concat", 00:10:16.183 "superblock": true, 00:10:16.183 "num_base_bdevs": 4, 00:10:16.183 "num_base_bdevs_discovered": 1, 00:10:16.183 "num_base_bdevs_operational": 4, 00:10:16.183 "base_bdevs_list": [ 00:10:16.183 { 00:10:16.183 "name": "BaseBdev1", 00:10:16.183 "uuid": "7003e355-7be5-49fa-8fb6-e6e7972f63fa", 00:10:16.183 "is_configured": true, 00:10:16.183 "data_offset": 2048, 00:10:16.183 "data_size": 63488 00:10:16.183 }, 00:10:16.183 { 00:10:16.183 "name": "BaseBdev2", 00:10:16.183 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:16.183 "is_configured": false, 00:10:16.183 "data_offset": 0, 00:10:16.183 "data_size": 0 00:10:16.183 }, 00:10:16.183 { 00:10:16.183 "name": "BaseBdev3", 00:10:16.183 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:16.183 "is_configured": false, 00:10:16.183 "data_offset": 0, 00:10:16.183 "data_size": 0 00:10:16.183 }, 00:10:16.183 { 00:10:16.183 "name": "BaseBdev4", 00:10:16.183 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:16.183 "is_configured": false, 00:10:16.183 "data_offset": 0, 00:10:16.183 "data_size": 0 00:10:16.183 } 00:10:16.183 ] 00:10:16.183 }' 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:16.183 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.441 [2024-12-05 19:00:33.922783] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:16.441 [2024-12-05 19:00:33.922869] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.441 [2024-12-05 19:00:33.934843] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:16.441 [2024-12-05 19:00:33.936754] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:16.441 [2024-12-05 19:00:33.936826] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:16.441 [2024-12-05 19:00:33.936860] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:16.441 [2024-12-05 19:00:33.936883] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:16.441 [2024-12-05 19:00:33.936931] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:16.441 [2024-12-05 19:00:33.936952] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:16.441 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:16.442 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:16.442 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:16.442 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:16.442 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.442 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:16.442 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.442 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:16.442 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:16.442 "name": "Existed_Raid", 00:10:16.442 "uuid": "dd7df643-13a0-46ca-8812-0a9399021301", 00:10:16.442 "strip_size_kb": 64, 00:10:16.442 "state": "configuring", 00:10:16.442 "raid_level": "concat", 00:10:16.442 "superblock": true, 00:10:16.442 "num_base_bdevs": 4, 00:10:16.442 "num_base_bdevs_discovered": 1, 00:10:16.442 "num_base_bdevs_operational": 4, 00:10:16.442 "base_bdevs_list": [ 00:10:16.442 { 00:10:16.442 "name": "BaseBdev1", 00:10:16.442 "uuid": "7003e355-7be5-49fa-8fb6-e6e7972f63fa", 00:10:16.442 "is_configured": true, 00:10:16.442 "data_offset": 2048, 00:10:16.442 "data_size": 63488 00:10:16.442 }, 00:10:16.442 { 00:10:16.442 "name": "BaseBdev2", 00:10:16.442 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:16.442 "is_configured": false, 00:10:16.442 "data_offset": 0, 00:10:16.442 "data_size": 0 00:10:16.442 }, 00:10:16.442 { 00:10:16.442 "name": "BaseBdev3", 00:10:16.442 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:16.442 "is_configured": false, 00:10:16.442 "data_offset": 0, 00:10:16.442 "data_size": 0 00:10:16.442 }, 00:10:16.442 { 00:10:16.442 "name": "BaseBdev4", 00:10:16.442 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:16.442 "is_configured": false, 00:10:16.442 "data_offset": 0, 00:10:16.442 "data_size": 0 00:10:16.442 } 00:10:16.442 ] 00:10:16.442 }' 00:10:16.442 19:00:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:16.442 19:00:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.009 [2024-12-05 19:00:34.293278] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:17.009 BaseBdev2 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.009 [ 00:10:17.009 { 00:10:17.009 "name": "BaseBdev2", 00:10:17.009 "aliases": [ 00:10:17.009 "31e48fa1-d2e4-42ab-b3c0-b1160f95500c" 00:10:17.009 ], 00:10:17.009 "product_name": "Malloc disk", 00:10:17.009 "block_size": 512, 00:10:17.009 "num_blocks": 65536, 00:10:17.009 "uuid": "31e48fa1-d2e4-42ab-b3c0-b1160f95500c", 00:10:17.009 "assigned_rate_limits": { 00:10:17.009 "rw_ios_per_sec": 0, 00:10:17.009 "rw_mbytes_per_sec": 0, 00:10:17.009 "r_mbytes_per_sec": 0, 00:10:17.009 "w_mbytes_per_sec": 0 00:10:17.009 }, 00:10:17.009 "claimed": true, 00:10:17.009 "claim_type": "exclusive_write", 00:10:17.009 "zoned": false, 00:10:17.009 "supported_io_types": { 00:10:17.009 "read": true, 00:10:17.009 "write": true, 00:10:17.009 "unmap": true, 00:10:17.009 "flush": true, 00:10:17.009 "reset": true, 00:10:17.009 "nvme_admin": false, 00:10:17.009 "nvme_io": false, 00:10:17.009 "nvme_io_md": false, 00:10:17.009 "write_zeroes": true, 00:10:17.009 "zcopy": true, 00:10:17.009 "get_zone_info": false, 00:10:17.009 "zone_management": false, 00:10:17.009 "zone_append": false, 00:10:17.009 "compare": false, 00:10:17.009 "compare_and_write": false, 00:10:17.009 "abort": true, 00:10:17.009 "seek_hole": false, 00:10:17.009 "seek_data": false, 00:10:17.009 "copy": true, 00:10:17.009 "nvme_iov_md": false 00:10:17.009 }, 00:10:17.009 "memory_domains": [ 00:10:17.009 { 00:10:17.009 "dma_device_id": "system", 00:10:17.009 "dma_device_type": 1 00:10:17.009 }, 00:10:17.009 { 00:10:17.009 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.009 "dma_device_type": 2 00:10:17.009 } 00:10:17.009 ], 00:10:17.009 "driver_specific": {} 00:10:17.009 } 00:10:17.009 ] 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.009 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:17.009 "name": "Existed_Raid", 00:10:17.009 "uuid": "dd7df643-13a0-46ca-8812-0a9399021301", 00:10:17.009 "strip_size_kb": 64, 00:10:17.009 "state": "configuring", 00:10:17.009 "raid_level": "concat", 00:10:17.009 "superblock": true, 00:10:17.009 "num_base_bdevs": 4, 00:10:17.009 "num_base_bdevs_discovered": 2, 00:10:17.009 "num_base_bdevs_operational": 4, 00:10:17.010 "base_bdevs_list": [ 00:10:17.010 { 00:10:17.010 "name": "BaseBdev1", 00:10:17.010 "uuid": "7003e355-7be5-49fa-8fb6-e6e7972f63fa", 00:10:17.010 "is_configured": true, 00:10:17.010 "data_offset": 2048, 00:10:17.010 "data_size": 63488 00:10:17.010 }, 00:10:17.010 { 00:10:17.010 "name": "BaseBdev2", 00:10:17.010 "uuid": "31e48fa1-d2e4-42ab-b3c0-b1160f95500c", 00:10:17.010 "is_configured": true, 00:10:17.010 "data_offset": 2048, 00:10:17.010 "data_size": 63488 00:10:17.010 }, 00:10:17.010 { 00:10:17.010 "name": "BaseBdev3", 00:10:17.010 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:17.010 "is_configured": false, 00:10:17.010 "data_offset": 0, 00:10:17.010 "data_size": 0 00:10:17.010 }, 00:10:17.010 { 00:10:17.010 "name": "BaseBdev4", 00:10:17.010 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:17.010 "is_configured": false, 00:10:17.010 "data_offset": 0, 00:10:17.010 "data_size": 0 00:10:17.010 } 00:10:17.010 ] 00:10:17.010 }' 00:10:17.010 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:17.010 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.268 [2024-12-05 19:00:34.740126] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:17.268 BaseBdev3 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.268 [ 00:10:17.268 { 00:10:17.268 "name": "BaseBdev3", 00:10:17.268 "aliases": [ 00:10:17.268 "9e59b752-b5a2-454e-941b-3c934cc5fc30" 00:10:17.268 ], 00:10:17.268 "product_name": "Malloc disk", 00:10:17.268 "block_size": 512, 00:10:17.268 "num_blocks": 65536, 00:10:17.268 "uuid": "9e59b752-b5a2-454e-941b-3c934cc5fc30", 00:10:17.268 "assigned_rate_limits": { 00:10:17.268 "rw_ios_per_sec": 0, 00:10:17.268 "rw_mbytes_per_sec": 0, 00:10:17.268 "r_mbytes_per_sec": 0, 00:10:17.268 "w_mbytes_per_sec": 0 00:10:17.268 }, 00:10:17.268 "claimed": true, 00:10:17.268 "claim_type": "exclusive_write", 00:10:17.268 "zoned": false, 00:10:17.268 "supported_io_types": { 00:10:17.268 "read": true, 00:10:17.268 "write": true, 00:10:17.268 "unmap": true, 00:10:17.268 "flush": true, 00:10:17.268 "reset": true, 00:10:17.268 "nvme_admin": false, 00:10:17.268 "nvme_io": false, 00:10:17.268 "nvme_io_md": false, 00:10:17.268 "write_zeroes": true, 00:10:17.268 "zcopy": true, 00:10:17.268 "get_zone_info": false, 00:10:17.268 "zone_management": false, 00:10:17.268 "zone_append": false, 00:10:17.268 "compare": false, 00:10:17.268 "compare_and_write": false, 00:10:17.268 "abort": true, 00:10:17.268 "seek_hole": false, 00:10:17.268 "seek_data": false, 00:10:17.268 "copy": true, 00:10:17.268 "nvme_iov_md": false 00:10:17.268 }, 00:10:17.268 "memory_domains": [ 00:10:17.268 { 00:10:17.268 "dma_device_id": "system", 00:10:17.268 "dma_device_type": 1 00:10:17.268 }, 00:10:17.268 { 00:10:17.268 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.268 "dma_device_type": 2 00:10:17.268 } 00:10:17.268 ], 00:10:17.268 "driver_specific": {} 00:10:17.268 } 00:10:17.268 ] 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.268 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.269 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.269 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:17.269 "name": "Existed_Raid", 00:10:17.269 "uuid": "dd7df643-13a0-46ca-8812-0a9399021301", 00:10:17.269 "strip_size_kb": 64, 00:10:17.269 "state": "configuring", 00:10:17.269 "raid_level": "concat", 00:10:17.269 "superblock": true, 00:10:17.269 "num_base_bdevs": 4, 00:10:17.269 "num_base_bdevs_discovered": 3, 00:10:17.269 "num_base_bdevs_operational": 4, 00:10:17.269 "base_bdevs_list": [ 00:10:17.269 { 00:10:17.269 "name": "BaseBdev1", 00:10:17.269 "uuid": "7003e355-7be5-49fa-8fb6-e6e7972f63fa", 00:10:17.269 "is_configured": true, 00:10:17.269 "data_offset": 2048, 00:10:17.269 "data_size": 63488 00:10:17.269 }, 00:10:17.269 { 00:10:17.269 "name": "BaseBdev2", 00:10:17.269 "uuid": "31e48fa1-d2e4-42ab-b3c0-b1160f95500c", 00:10:17.269 "is_configured": true, 00:10:17.269 "data_offset": 2048, 00:10:17.269 "data_size": 63488 00:10:17.269 }, 00:10:17.269 { 00:10:17.269 "name": "BaseBdev3", 00:10:17.269 "uuid": "9e59b752-b5a2-454e-941b-3c934cc5fc30", 00:10:17.269 "is_configured": true, 00:10:17.269 "data_offset": 2048, 00:10:17.269 "data_size": 63488 00:10:17.269 }, 00:10:17.269 { 00:10:17.269 "name": "BaseBdev4", 00:10:17.269 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:17.269 "is_configured": false, 00:10:17.269 "data_offset": 0, 00:10:17.269 "data_size": 0 00:10:17.269 } 00:10:17.269 ] 00:10:17.269 }' 00:10:17.269 19:00:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:17.269 19:00:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.835 [2024-12-05 19:00:35.198466] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:17.835 [2024-12-05 19:00:35.198684] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:17.835 [2024-12-05 19:00:35.198702] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:17.835 BaseBdev4 00:10:17.835 [2024-12-05 19:00:35.199007] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:17.835 [2024-12-05 19:00:35.199147] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:17.835 [2024-12-05 19:00:35.199165] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:17.835 [2024-12-05 19:00:35.199312] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.835 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.835 [ 00:10:17.835 { 00:10:17.835 "name": "BaseBdev4", 00:10:17.835 "aliases": [ 00:10:17.835 "8aba96f5-aefd-4b30-86da-e41c3513ba54" 00:10:17.835 ], 00:10:17.835 "product_name": "Malloc disk", 00:10:17.835 "block_size": 512, 00:10:17.835 "num_blocks": 65536, 00:10:17.835 "uuid": "8aba96f5-aefd-4b30-86da-e41c3513ba54", 00:10:17.835 "assigned_rate_limits": { 00:10:17.835 "rw_ios_per_sec": 0, 00:10:17.835 "rw_mbytes_per_sec": 0, 00:10:17.835 "r_mbytes_per_sec": 0, 00:10:17.835 "w_mbytes_per_sec": 0 00:10:17.835 }, 00:10:17.835 "claimed": true, 00:10:17.835 "claim_type": "exclusive_write", 00:10:17.835 "zoned": false, 00:10:17.835 "supported_io_types": { 00:10:17.835 "read": true, 00:10:17.835 "write": true, 00:10:17.835 "unmap": true, 00:10:17.835 "flush": true, 00:10:17.835 "reset": true, 00:10:17.835 "nvme_admin": false, 00:10:17.835 "nvme_io": false, 00:10:17.835 "nvme_io_md": false, 00:10:17.835 "write_zeroes": true, 00:10:17.835 "zcopy": true, 00:10:17.835 "get_zone_info": false, 00:10:17.836 "zone_management": false, 00:10:17.836 "zone_append": false, 00:10:17.836 "compare": false, 00:10:17.836 "compare_and_write": false, 00:10:17.836 "abort": true, 00:10:17.836 "seek_hole": false, 00:10:17.836 "seek_data": false, 00:10:17.836 "copy": true, 00:10:17.836 "nvme_iov_md": false 00:10:17.836 }, 00:10:17.836 "memory_domains": [ 00:10:17.836 { 00:10:17.836 "dma_device_id": "system", 00:10:17.836 "dma_device_type": 1 00:10:17.836 }, 00:10:17.836 { 00:10:17.836 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.836 "dma_device_type": 2 00:10:17.836 } 00:10:17.836 ], 00:10:17.836 "driver_specific": {} 00:10:17.836 } 00:10:17.836 ] 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:17.836 "name": "Existed_Raid", 00:10:17.836 "uuid": "dd7df643-13a0-46ca-8812-0a9399021301", 00:10:17.836 "strip_size_kb": 64, 00:10:17.836 "state": "online", 00:10:17.836 "raid_level": "concat", 00:10:17.836 "superblock": true, 00:10:17.836 "num_base_bdevs": 4, 00:10:17.836 "num_base_bdevs_discovered": 4, 00:10:17.836 "num_base_bdevs_operational": 4, 00:10:17.836 "base_bdevs_list": [ 00:10:17.836 { 00:10:17.836 "name": "BaseBdev1", 00:10:17.836 "uuid": "7003e355-7be5-49fa-8fb6-e6e7972f63fa", 00:10:17.836 "is_configured": true, 00:10:17.836 "data_offset": 2048, 00:10:17.836 "data_size": 63488 00:10:17.836 }, 00:10:17.836 { 00:10:17.836 "name": "BaseBdev2", 00:10:17.836 "uuid": "31e48fa1-d2e4-42ab-b3c0-b1160f95500c", 00:10:17.836 "is_configured": true, 00:10:17.836 "data_offset": 2048, 00:10:17.836 "data_size": 63488 00:10:17.836 }, 00:10:17.836 { 00:10:17.836 "name": "BaseBdev3", 00:10:17.836 "uuid": "9e59b752-b5a2-454e-941b-3c934cc5fc30", 00:10:17.836 "is_configured": true, 00:10:17.836 "data_offset": 2048, 00:10:17.836 "data_size": 63488 00:10:17.836 }, 00:10:17.836 { 00:10:17.836 "name": "BaseBdev4", 00:10:17.836 "uuid": "8aba96f5-aefd-4b30-86da-e41c3513ba54", 00:10:17.836 "is_configured": true, 00:10:17.836 "data_offset": 2048, 00:10:17.836 "data_size": 63488 00:10:17.836 } 00:10:17.836 ] 00:10:17.836 }' 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:17.836 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.094 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:18.094 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:18.094 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:18.094 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:18.094 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:18.094 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:18.094 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:18.094 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.094 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.094 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:18.094 [2024-12-05 19:00:35.650037] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:18.352 "name": "Existed_Raid", 00:10:18.352 "aliases": [ 00:10:18.352 "dd7df643-13a0-46ca-8812-0a9399021301" 00:10:18.352 ], 00:10:18.352 "product_name": "Raid Volume", 00:10:18.352 "block_size": 512, 00:10:18.352 "num_blocks": 253952, 00:10:18.352 "uuid": "dd7df643-13a0-46ca-8812-0a9399021301", 00:10:18.352 "assigned_rate_limits": { 00:10:18.352 "rw_ios_per_sec": 0, 00:10:18.352 "rw_mbytes_per_sec": 0, 00:10:18.352 "r_mbytes_per_sec": 0, 00:10:18.352 "w_mbytes_per_sec": 0 00:10:18.352 }, 00:10:18.352 "claimed": false, 00:10:18.352 "zoned": false, 00:10:18.352 "supported_io_types": { 00:10:18.352 "read": true, 00:10:18.352 "write": true, 00:10:18.352 "unmap": true, 00:10:18.352 "flush": true, 00:10:18.352 "reset": true, 00:10:18.352 "nvme_admin": false, 00:10:18.352 "nvme_io": false, 00:10:18.352 "nvme_io_md": false, 00:10:18.352 "write_zeroes": true, 00:10:18.352 "zcopy": false, 00:10:18.352 "get_zone_info": false, 00:10:18.352 "zone_management": false, 00:10:18.352 "zone_append": false, 00:10:18.352 "compare": false, 00:10:18.352 "compare_and_write": false, 00:10:18.352 "abort": false, 00:10:18.352 "seek_hole": false, 00:10:18.352 "seek_data": false, 00:10:18.352 "copy": false, 00:10:18.352 "nvme_iov_md": false 00:10:18.352 }, 00:10:18.352 "memory_domains": [ 00:10:18.352 { 00:10:18.352 "dma_device_id": "system", 00:10:18.352 "dma_device_type": 1 00:10:18.352 }, 00:10:18.352 { 00:10:18.352 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:18.352 "dma_device_type": 2 00:10:18.352 }, 00:10:18.352 { 00:10:18.352 "dma_device_id": "system", 00:10:18.352 "dma_device_type": 1 00:10:18.352 }, 00:10:18.352 { 00:10:18.352 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:18.352 "dma_device_type": 2 00:10:18.352 }, 00:10:18.352 { 00:10:18.352 "dma_device_id": "system", 00:10:18.352 "dma_device_type": 1 00:10:18.352 }, 00:10:18.352 { 00:10:18.352 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:18.352 "dma_device_type": 2 00:10:18.352 }, 00:10:18.352 { 00:10:18.352 "dma_device_id": "system", 00:10:18.352 "dma_device_type": 1 00:10:18.352 }, 00:10:18.352 { 00:10:18.352 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:18.352 "dma_device_type": 2 00:10:18.352 } 00:10:18.352 ], 00:10:18.352 "driver_specific": { 00:10:18.352 "raid": { 00:10:18.352 "uuid": "dd7df643-13a0-46ca-8812-0a9399021301", 00:10:18.352 "strip_size_kb": 64, 00:10:18.352 "state": "online", 00:10:18.352 "raid_level": "concat", 00:10:18.352 "superblock": true, 00:10:18.352 "num_base_bdevs": 4, 00:10:18.352 "num_base_bdevs_discovered": 4, 00:10:18.352 "num_base_bdevs_operational": 4, 00:10:18.352 "base_bdevs_list": [ 00:10:18.352 { 00:10:18.352 "name": "BaseBdev1", 00:10:18.352 "uuid": "7003e355-7be5-49fa-8fb6-e6e7972f63fa", 00:10:18.352 "is_configured": true, 00:10:18.352 "data_offset": 2048, 00:10:18.352 "data_size": 63488 00:10:18.352 }, 00:10:18.352 { 00:10:18.352 "name": "BaseBdev2", 00:10:18.352 "uuid": "31e48fa1-d2e4-42ab-b3c0-b1160f95500c", 00:10:18.352 "is_configured": true, 00:10:18.352 "data_offset": 2048, 00:10:18.352 "data_size": 63488 00:10:18.352 }, 00:10:18.352 { 00:10:18.352 "name": "BaseBdev3", 00:10:18.352 "uuid": "9e59b752-b5a2-454e-941b-3c934cc5fc30", 00:10:18.352 "is_configured": true, 00:10:18.352 "data_offset": 2048, 00:10:18.352 "data_size": 63488 00:10:18.352 }, 00:10:18.352 { 00:10:18.352 "name": "BaseBdev4", 00:10:18.352 "uuid": "8aba96f5-aefd-4b30-86da-e41c3513ba54", 00:10:18.352 "is_configured": true, 00:10:18.352 "data_offset": 2048, 00:10:18.352 "data_size": 63488 00:10:18.352 } 00:10:18.352 ] 00:10:18.352 } 00:10:18.352 } 00:10:18.352 }' 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:18.352 BaseBdev2 00:10:18.352 BaseBdev3 00:10:18.352 BaseBdev4' 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.352 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.612 [2024-12-05 19:00:35.937276] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:18.612 [2024-12-05 19:00:35.937350] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:18.612 [2024-12-05 19:00:35.937437] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:18.612 "name": "Existed_Raid", 00:10:18.612 "uuid": "dd7df643-13a0-46ca-8812-0a9399021301", 00:10:18.612 "strip_size_kb": 64, 00:10:18.612 "state": "offline", 00:10:18.612 "raid_level": "concat", 00:10:18.612 "superblock": true, 00:10:18.612 "num_base_bdevs": 4, 00:10:18.612 "num_base_bdevs_discovered": 3, 00:10:18.612 "num_base_bdevs_operational": 3, 00:10:18.612 "base_bdevs_list": [ 00:10:18.612 { 00:10:18.612 "name": null, 00:10:18.612 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:18.612 "is_configured": false, 00:10:18.612 "data_offset": 0, 00:10:18.612 "data_size": 63488 00:10:18.612 }, 00:10:18.612 { 00:10:18.612 "name": "BaseBdev2", 00:10:18.612 "uuid": "31e48fa1-d2e4-42ab-b3c0-b1160f95500c", 00:10:18.612 "is_configured": true, 00:10:18.612 "data_offset": 2048, 00:10:18.612 "data_size": 63488 00:10:18.612 }, 00:10:18.612 { 00:10:18.612 "name": "BaseBdev3", 00:10:18.612 "uuid": "9e59b752-b5a2-454e-941b-3c934cc5fc30", 00:10:18.612 "is_configured": true, 00:10:18.612 "data_offset": 2048, 00:10:18.612 "data_size": 63488 00:10:18.612 }, 00:10:18.612 { 00:10:18.612 "name": "BaseBdev4", 00:10:18.612 "uuid": "8aba96f5-aefd-4b30-86da-e41c3513ba54", 00:10:18.612 "is_configured": true, 00:10:18.612 "data_offset": 2048, 00:10:18.612 "data_size": 63488 00:10:18.612 } 00:10:18.612 ] 00:10:18.612 }' 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:18.612 19:00:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.870 [2024-12-05 19:00:36.343984] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:18.870 [2024-12-05 19:00:36.414977] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:18.870 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:19.129 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:19.129 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.130 [2024-12-05 19:00:36.465748] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:19.130 [2024-12-05 19:00:36.465834] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.130 BaseBdev2 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.130 [ 00:10:19.130 { 00:10:19.130 "name": "BaseBdev2", 00:10:19.130 "aliases": [ 00:10:19.130 "3e1433a9-d073-4e50-8bef-d72b5541329e" 00:10:19.130 ], 00:10:19.130 "product_name": "Malloc disk", 00:10:19.130 "block_size": 512, 00:10:19.130 "num_blocks": 65536, 00:10:19.130 "uuid": "3e1433a9-d073-4e50-8bef-d72b5541329e", 00:10:19.130 "assigned_rate_limits": { 00:10:19.130 "rw_ios_per_sec": 0, 00:10:19.130 "rw_mbytes_per_sec": 0, 00:10:19.130 "r_mbytes_per_sec": 0, 00:10:19.130 "w_mbytes_per_sec": 0 00:10:19.130 }, 00:10:19.130 "claimed": false, 00:10:19.130 "zoned": false, 00:10:19.130 "supported_io_types": { 00:10:19.130 "read": true, 00:10:19.130 "write": true, 00:10:19.130 "unmap": true, 00:10:19.130 "flush": true, 00:10:19.130 "reset": true, 00:10:19.130 "nvme_admin": false, 00:10:19.130 "nvme_io": false, 00:10:19.130 "nvme_io_md": false, 00:10:19.130 "write_zeroes": true, 00:10:19.130 "zcopy": true, 00:10:19.130 "get_zone_info": false, 00:10:19.130 "zone_management": false, 00:10:19.130 "zone_append": false, 00:10:19.130 "compare": false, 00:10:19.130 "compare_and_write": false, 00:10:19.130 "abort": true, 00:10:19.130 "seek_hole": false, 00:10:19.130 "seek_data": false, 00:10:19.130 "copy": true, 00:10:19.130 "nvme_iov_md": false 00:10:19.130 }, 00:10:19.130 "memory_domains": [ 00:10:19.130 { 00:10:19.130 "dma_device_id": "system", 00:10:19.130 "dma_device_type": 1 00:10:19.130 }, 00:10:19.130 { 00:10:19.130 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:19.130 "dma_device_type": 2 00:10:19.130 } 00:10:19.130 ], 00:10:19.130 "driver_specific": {} 00:10:19.130 } 00:10:19.130 ] 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.130 BaseBdev3 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.130 [ 00:10:19.130 { 00:10:19.130 "name": "BaseBdev3", 00:10:19.130 "aliases": [ 00:10:19.130 "dfb40bf8-0670-4bdf-8f4b-83836990c11d" 00:10:19.130 ], 00:10:19.130 "product_name": "Malloc disk", 00:10:19.130 "block_size": 512, 00:10:19.130 "num_blocks": 65536, 00:10:19.130 "uuid": "dfb40bf8-0670-4bdf-8f4b-83836990c11d", 00:10:19.130 "assigned_rate_limits": { 00:10:19.130 "rw_ios_per_sec": 0, 00:10:19.130 "rw_mbytes_per_sec": 0, 00:10:19.130 "r_mbytes_per_sec": 0, 00:10:19.130 "w_mbytes_per_sec": 0 00:10:19.130 }, 00:10:19.130 "claimed": false, 00:10:19.130 "zoned": false, 00:10:19.130 "supported_io_types": { 00:10:19.130 "read": true, 00:10:19.130 "write": true, 00:10:19.130 "unmap": true, 00:10:19.130 "flush": true, 00:10:19.130 "reset": true, 00:10:19.130 "nvme_admin": false, 00:10:19.130 "nvme_io": false, 00:10:19.130 "nvme_io_md": false, 00:10:19.130 "write_zeroes": true, 00:10:19.130 "zcopy": true, 00:10:19.130 "get_zone_info": false, 00:10:19.130 "zone_management": false, 00:10:19.130 "zone_append": false, 00:10:19.130 "compare": false, 00:10:19.130 "compare_and_write": false, 00:10:19.130 "abort": true, 00:10:19.130 "seek_hole": false, 00:10:19.130 "seek_data": false, 00:10:19.130 "copy": true, 00:10:19.130 "nvme_iov_md": false 00:10:19.130 }, 00:10:19.130 "memory_domains": [ 00:10:19.130 { 00:10:19.130 "dma_device_id": "system", 00:10:19.130 "dma_device_type": 1 00:10:19.130 }, 00:10:19.130 { 00:10:19.130 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:19.130 "dma_device_type": 2 00:10:19.130 } 00:10:19.130 ], 00:10:19.130 "driver_specific": {} 00:10:19.130 } 00:10:19.130 ] 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.130 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.131 BaseBdev4 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.131 [ 00:10:19.131 { 00:10:19.131 "name": "BaseBdev4", 00:10:19.131 "aliases": [ 00:10:19.131 "bd9be15d-6afe-4951-8b3c-aac2b29119f4" 00:10:19.131 ], 00:10:19.131 "product_name": "Malloc disk", 00:10:19.131 "block_size": 512, 00:10:19.131 "num_blocks": 65536, 00:10:19.131 "uuid": "bd9be15d-6afe-4951-8b3c-aac2b29119f4", 00:10:19.131 "assigned_rate_limits": { 00:10:19.131 "rw_ios_per_sec": 0, 00:10:19.131 "rw_mbytes_per_sec": 0, 00:10:19.131 "r_mbytes_per_sec": 0, 00:10:19.131 "w_mbytes_per_sec": 0 00:10:19.131 }, 00:10:19.131 "claimed": false, 00:10:19.131 "zoned": false, 00:10:19.131 "supported_io_types": { 00:10:19.131 "read": true, 00:10:19.131 "write": true, 00:10:19.131 "unmap": true, 00:10:19.131 "flush": true, 00:10:19.131 "reset": true, 00:10:19.131 "nvme_admin": false, 00:10:19.131 "nvme_io": false, 00:10:19.131 "nvme_io_md": false, 00:10:19.131 "write_zeroes": true, 00:10:19.131 "zcopy": true, 00:10:19.131 "get_zone_info": false, 00:10:19.131 "zone_management": false, 00:10:19.131 "zone_append": false, 00:10:19.131 "compare": false, 00:10:19.131 "compare_and_write": false, 00:10:19.131 "abort": true, 00:10:19.131 "seek_hole": false, 00:10:19.131 "seek_data": false, 00:10:19.131 "copy": true, 00:10:19.131 "nvme_iov_md": false 00:10:19.131 }, 00:10:19.131 "memory_domains": [ 00:10:19.131 { 00:10:19.131 "dma_device_id": "system", 00:10:19.131 "dma_device_type": 1 00:10:19.131 }, 00:10:19.131 { 00:10:19.131 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:19.131 "dma_device_type": 2 00:10:19.131 } 00:10:19.131 ], 00:10:19.131 "driver_specific": {} 00:10:19.131 } 00:10:19.131 ] 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:19.131 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:19.390 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.391 [2024-12-05 19:00:36.693429] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:19.391 [2024-12-05 19:00:36.693532] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:19.391 [2024-12-05 19:00:36.693594] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:19.391 [2024-12-05 19:00:36.695397] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:19.391 [2024-12-05 19:00:36.695485] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:19.391 "name": "Existed_Raid", 00:10:19.391 "uuid": "0d373c1f-8a94-4359-a74d-897c99420b35", 00:10:19.391 "strip_size_kb": 64, 00:10:19.391 "state": "configuring", 00:10:19.391 "raid_level": "concat", 00:10:19.391 "superblock": true, 00:10:19.391 "num_base_bdevs": 4, 00:10:19.391 "num_base_bdevs_discovered": 3, 00:10:19.391 "num_base_bdevs_operational": 4, 00:10:19.391 "base_bdevs_list": [ 00:10:19.391 { 00:10:19.391 "name": "BaseBdev1", 00:10:19.391 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:19.391 "is_configured": false, 00:10:19.391 "data_offset": 0, 00:10:19.391 "data_size": 0 00:10:19.391 }, 00:10:19.391 { 00:10:19.391 "name": "BaseBdev2", 00:10:19.391 "uuid": "3e1433a9-d073-4e50-8bef-d72b5541329e", 00:10:19.391 "is_configured": true, 00:10:19.391 "data_offset": 2048, 00:10:19.391 "data_size": 63488 00:10:19.391 }, 00:10:19.391 { 00:10:19.391 "name": "BaseBdev3", 00:10:19.391 "uuid": "dfb40bf8-0670-4bdf-8f4b-83836990c11d", 00:10:19.391 "is_configured": true, 00:10:19.391 "data_offset": 2048, 00:10:19.391 "data_size": 63488 00:10:19.391 }, 00:10:19.391 { 00:10:19.391 "name": "BaseBdev4", 00:10:19.391 "uuid": "bd9be15d-6afe-4951-8b3c-aac2b29119f4", 00:10:19.391 "is_configured": true, 00:10:19.391 "data_offset": 2048, 00:10:19.391 "data_size": 63488 00:10:19.391 } 00:10:19.391 ] 00:10:19.391 }' 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:19.391 19:00:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.650 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:19.650 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.650 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.650 [2024-12-05 19:00:37.096709] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:19.651 "name": "Existed_Raid", 00:10:19.651 "uuid": "0d373c1f-8a94-4359-a74d-897c99420b35", 00:10:19.651 "strip_size_kb": 64, 00:10:19.651 "state": "configuring", 00:10:19.651 "raid_level": "concat", 00:10:19.651 "superblock": true, 00:10:19.651 "num_base_bdevs": 4, 00:10:19.651 "num_base_bdevs_discovered": 2, 00:10:19.651 "num_base_bdevs_operational": 4, 00:10:19.651 "base_bdevs_list": [ 00:10:19.651 { 00:10:19.651 "name": "BaseBdev1", 00:10:19.651 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:19.651 "is_configured": false, 00:10:19.651 "data_offset": 0, 00:10:19.651 "data_size": 0 00:10:19.651 }, 00:10:19.651 { 00:10:19.651 "name": null, 00:10:19.651 "uuid": "3e1433a9-d073-4e50-8bef-d72b5541329e", 00:10:19.651 "is_configured": false, 00:10:19.651 "data_offset": 0, 00:10:19.651 "data_size": 63488 00:10:19.651 }, 00:10:19.651 { 00:10:19.651 "name": "BaseBdev3", 00:10:19.651 "uuid": "dfb40bf8-0670-4bdf-8f4b-83836990c11d", 00:10:19.651 "is_configured": true, 00:10:19.651 "data_offset": 2048, 00:10:19.651 "data_size": 63488 00:10:19.651 }, 00:10:19.651 { 00:10:19.651 "name": "BaseBdev4", 00:10:19.651 "uuid": "bd9be15d-6afe-4951-8b3c-aac2b29119f4", 00:10:19.651 "is_configured": true, 00:10:19.651 "data_offset": 2048, 00:10:19.651 "data_size": 63488 00:10:19.651 } 00:10:19.651 ] 00:10:19.651 }' 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:19.651 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.254 [2024-12-05 19:00:37.570956] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:20.254 BaseBdev1 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.254 [ 00:10:20.254 { 00:10:20.254 "name": "BaseBdev1", 00:10:20.254 "aliases": [ 00:10:20.254 "cfcc4628-d121-4024-b9b1-d14276d72b97" 00:10:20.254 ], 00:10:20.254 "product_name": "Malloc disk", 00:10:20.254 "block_size": 512, 00:10:20.254 "num_blocks": 65536, 00:10:20.254 "uuid": "cfcc4628-d121-4024-b9b1-d14276d72b97", 00:10:20.254 "assigned_rate_limits": { 00:10:20.254 "rw_ios_per_sec": 0, 00:10:20.254 "rw_mbytes_per_sec": 0, 00:10:20.254 "r_mbytes_per_sec": 0, 00:10:20.254 "w_mbytes_per_sec": 0 00:10:20.254 }, 00:10:20.254 "claimed": true, 00:10:20.254 "claim_type": "exclusive_write", 00:10:20.254 "zoned": false, 00:10:20.254 "supported_io_types": { 00:10:20.254 "read": true, 00:10:20.254 "write": true, 00:10:20.254 "unmap": true, 00:10:20.254 "flush": true, 00:10:20.254 "reset": true, 00:10:20.254 "nvme_admin": false, 00:10:20.254 "nvme_io": false, 00:10:20.254 "nvme_io_md": false, 00:10:20.254 "write_zeroes": true, 00:10:20.254 "zcopy": true, 00:10:20.254 "get_zone_info": false, 00:10:20.254 "zone_management": false, 00:10:20.254 "zone_append": false, 00:10:20.254 "compare": false, 00:10:20.254 "compare_and_write": false, 00:10:20.254 "abort": true, 00:10:20.254 "seek_hole": false, 00:10:20.254 "seek_data": false, 00:10:20.254 "copy": true, 00:10:20.254 "nvme_iov_md": false 00:10:20.254 }, 00:10:20.254 "memory_domains": [ 00:10:20.254 { 00:10:20.254 "dma_device_id": "system", 00:10:20.254 "dma_device_type": 1 00:10:20.254 }, 00:10:20.254 { 00:10:20.254 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:20.254 "dma_device_type": 2 00:10:20.254 } 00:10:20.254 ], 00:10:20.254 "driver_specific": {} 00:10:20.254 } 00:10:20.254 ] 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:20.254 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:20.255 "name": "Existed_Raid", 00:10:20.255 "uuid": "0d373c1f-8a94-4359-a74d-897c99420b35", 00:10:20.255 "strip_size_kb": 64, 00:10:20.255 "state": "configuring", 00:10:20.255 "raid_level": "concat", 00:10:20.255 "superblock": true, 00:10:20.255 "num_base_bdevs": 4, 00:10:20.255 "num_base_bdevs_discovered": 3, 00:10:20.255 "num_base_bdevs_operational": 4, 00:10:20.255 "base_bdevs_list": [ 00:10:20.255 { 00:10:20.255 "name": "BaseBdev1", 00:10:20.255 "uuid": "cfcc4628-d121-4024-b9b1-d14276d72b97", 00:10:20.255 "is_configured": true, 00:10:20.255 "data_offset": 2048, 00:10:20.255 "data_size": 63488 00:10:20.255 }, 00:10:20.255 { 00:10:20.255 "name": null, 00:10:20.255 "uuid": "3e1433a9-d073-4e50-8bef-d72b5541329e", 00:10:20.255 "is_configured": false, 00:10:20.255 "data_offset": 0, 00:10:20.255 "data_size": 63488 00:10:20.255 }, 00:10:20.255 { 00:10:20.255 "name": "BaseBdev3", 00:10:20.255 "uuid": "dfb40bf8-0670-4bdf-8f4b-83836990c11d", 00:10:20.255 "is_configured": true, 00:10:20.255 "data_offset": 2048, 00:10:20.255 "data_size": 63488 00:10:20.255 }, 00:10:20.255 { 00:10:20.255 "name": "BaseBdev4", 00:10:20.255 "uuid": "bd9be15d-6afe-4951-8b3c-aac2b29119f4", 00:10:20.255 "is_configured": true, 00:10:20.255 "data_offset": 2048, 00:10:20.255 "data_size": 63488 00:10:20.255 } 00:10:20.255 ] 00:10:20.255 }' 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:20.255 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.529 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.529 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.529 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.529 19:00:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:20.529 19:00:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.529 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:20.529 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:20.529 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.529 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.529 [2024-12-05 19:00:38.014234] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:20.529 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.529 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:20.529 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:20.529 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:20.529 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:20.529 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:20.530 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:20.530 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:20.530 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:20.530 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:20.530 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:20.530 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.530 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:20.530 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:20.530 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:20.530 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:20.530 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:20.530 "name": "Existed_Raid", 00:10:20.530 "uuid": "0d373c1f-8a94-4359-a74d-897c99420b35", 00:10:20.530 "strip_size_kb": 64, 00:10:20.530 "state": "configuring", 00:10:20.530 "raid_level": "concat", 00:10:20.530 "superblock": true, 00:10:20.530 "num_base_bdevs": 4, 00:10:20.530 "num_base_bdevs_discovered": 2, 00:10:20.530 "num_base_bdevs_operational": 4, 00:10:20.530 "base_bdevs_list": [ 00:10:20.530 { 00:10:20.530 "name": "BaseBdev1", 00:10:20.530 "uuid": "cfcc4628-d121-4024-b9b1-d14276d72b97", 00:10:20.530 "is_configured": true, 00:10:20.530 "data_offset": 2048, 00:10:20.530 "data_size": 63488 00:10:20.530 }, 00:10:20.530 { 00:10:20.530 "name": null, 00:10:20.530 "uuid": "3e1433a9-d073-4e50-8bef-d72b5541329e", 00:10:20.530 "is_configured": false, 00:10:20.530 "data_offset": 0, 00:10:20.530 "data_size": 63488 00:10:20.530 }, 00:10:20.530 { 00:10:20.530 "name": null, 00:10:20.530 "uuid": "dfb40bf8-0670-4bdf-8f4b-83836990c11d", 00:10:20.530 "is_configured": false, 00:10:20.530 "data_offset": 0, 00:10:20.530 "data_size": 63488 00:10:20.530 }, 00:10:20.530 { 00:10:20.530 "name": "BaseBdev4", 00:10:20.530 "uuid": "bd9be15d-6afe-4951-8b3c-aac2b29119f4", 00:10:20.530 "is_configured": true, 00:10:20.530 "data_offset": 2048, 00:10:20.530 "data_size": 63488 00:10:20.530 } 00:10:20.530 ] 00:10:20.530 }' 00:10:20.530 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:20.530 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.099 [2024-12-05 19:00:38.469514] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:21.099 "name": "Existed_Raid", 00:10:21.099 "uuid": "0d373c1f-8a94-4359-a74d-897c99420b35", 00:10:21.099 "strip_size_kb": 64, 00:10:21.099 "state": "configuring", 00:10:21.099 "raid_level": "concat", 00:10:21.099 "superblock": true, 00:10:21.099 "num_base_bdevs": 4, 00:10:21.099 "num_base_bdevs_discovered": 3, 00:10:21.099 "num_base_bdevs_operational": 4, 00:10:21.099 "base_bdevs_list": [ 00:10:21.099 { 00:10:21.099 "name": "BaseBdev1", 00:10:21.099 "uuid": "cfcc4628-d121-4024-b9b1-d14276d72b97", 00:10:21.099 "is_configured": true, 00:10:21.099 "data_offset": 2048, 00:10:21.099 "data_size": 63488 00:10:21.099 }, 00:10:21.099 { 00:10:21.099 "name": null, 00:10:21.099 "uuid": "3e1433a9-d073-4e50-8bef-d72b5541329e", 00:10:21.099 "is_configured": false, 00:10:21.099 "data_offset": 0, 00:10:21.099 "data_size": 63488 00:10:21.099 }, 00:10:21.099 { 00:10:21.099 "name": "BaseBdev3", 00:10:21.099 "uuid": "dfb40bf8-0670-4bdf-8f4b-83836990c11d", 00:10:21.099 "is_configured": true, 00:10:21.099 "data_offset": 2048, 00:10:21.099 "data_size": 63488 00:10:21.099 }, 00:10:21.099 { 00:10:21.099 "name": "BaseBdev4", 00:10:21.099 "uuid": "bd9be15d-6afe-4951-8b3c-aac2b29119f4", 00:10:21.099 "is_configured": true, 00:10:21.099 "data_offset": 2048, 00:10:21.099 "data_size": 63488 00:10:21.099 } 00:10:21.099 ] 00:10:21.099 }' 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:21.099 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.359 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:21.359 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.359 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.359 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.359 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.619 [2024-12-05 19:00:38.924761] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:21.619 "name": "Existed_Raid", 00:10:21.619 "uuid": "0d373c1f-8a94-4359-a74d-897c99420b35", 00:10:21.619 "strip_size_kb": 64, 00:10:21.619 "state": "configuring", 00:10:21.619 "raid_level": "concat", 00:10:21.619 "superblock": true, 00:10:21.619 "num_base_bdevs": 4, 00:10:21.619 "num_base_bdevs_discovered": 2, 00:10:21.619 "num_base_bdevs_operational": 4, 00:10:21.619 "base_bdevs_list": [ 00:10:21.619 { 00:10:21.619 "name": null, 00:10:21.619 "uuid": "cfcc4628-d121-4024-b9b1-d14276d72b97", 00:10:21.619 "is_configured": false, 00:10:21.619 "data_offset": 0, 00:10:21.619 "data_size": 63488 00:10:21.619 }, 00:10:21.619 { 00:10:21.619 "name": null, 00:10:21.619 "uuid": "3e1433a9-d073-4e50-8bef-d72b5541329e", 00:10:21.619 "is_configured": false, 00:10:21.619 "data_offset": 0, 00:10:21.619 "data_size": 63488 00:10:21.619 }, 00:10:21.619 { 00:10:21.619 "name": "BaseBdev3", 00:10:21.619 "uuid": "dfb40bf8-0670-4bdf-8f4b-83836990c11d", 00:10:21.619 "is_configured": true, 00:10:21.619 "data_offset": 2048, 00:10:21.619 "data_size": 63488 00:10:21.619 }, 00:10:21.619 { 00:10:21.619 "name": "BaseBdev4", 00:10:21.619 "uuid": "bd9be15d-6afe-4951-8b3c-aac2b29119f4", 00:10:21.619 "is_configured": true, 00:10:21.619 "data_offset": 2048, 00:10:21.619 "data_size": 63488 00:10:21.619 } 00:10:21.619 ] 00:10:21.619 }' 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:21.619 19:00:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:21.880 [2024-12-05 19:00:39.422407] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:21.880 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.140 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.140 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:22.141 "name": "Existed_Raid", 00:10:22.141 "uuid": "0d373c1f-8a94-4359-a74d-897c99420b35", 00:10:22.141 "strip_size_kb": 64, 00:10:22.141 "state": "configuring", 00:10:22.141 "raid_level": "concat", 00:10:22.141 "superblock": true, 00:10:22.141 "num_base_bdevs": 4, 00:10:22.141 "num_base_bdevs_discovered": 3, 00:10:22.141 "num_base_bdevs_operational": 4, 00:10:22.141 "base_bdevs_list": [ 00:10:22.141 { 00:10:22.141 "name": null, 00:10:22.141 "uuid": "cfcc4628-d121-4024-b9b1-d14276d72b97", 00:10:22.141 "is_configured": false, 00:10:22.141 "data_offset": 0, 00:10:22.141 "data_size": 63488 00:10:22.141 }, 00:10:22.141 { 00:10:22.141 "name": "BaseBdev2", 00:10:22.141 "uuid": "3e1433a9-d073-4e50-8bef-d72b5541329e", 00:10:22.141 "is_configured": true, 00:10:22.141 "data_offset": 2048, 00:10:22.141 "data_size": 63488 00:10:22.141 }, 00:10:22.141 { 00:10:22.141 "name": "BaseBdev3", 00:10:22.141 "uuid": "dfb40bf8-0670-4bdf-8f4b-83836990c11d", 00:10:22.141 "is_configured": true, 00:10:22.141 "data_offset": 2048, 00:10:22.141 "data_size": 63488 00:10:22.141 }, 00:10:22.141 { 00:10:22.141 "name": "BaseBdev4", 00:10:22.141 "uuid": "bd9be15d-6afe-4951-8b3c-aac2b29119f4", 00:10:22.141 "is_configured": true, 00:10:22.141 "data_offset": 2048, 00:10:22.141 "data_size": 63488 00:10:22.141 } 00:10:22.141 ] 00:10:22.141 }' 00:10:22.141 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:22.141 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u cfcc4628-d121-4024-b9b1-d14276d72b97 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.401 NewBaseBdev 00:10:22.401 [2024-12-05 19:00:39.840776] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:22.401 [2024-12-05 19:00:39.840955] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:22.401 [2024-12-05 19:00:39.840968] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:22.401 [2024-12-05 19:00:39.841238] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:22.401 [2024-12-05 19:00:39.841361] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:22.401 [2024-12-05 19:00:39.841378] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:22.401 [2024-12-05 19:00:39.841473] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.401 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.401 [ 00:10:22.401 { 00:10:22.401 "name": "NewBaseBdev", 00:10:22.401 "aliases": [ 00:10:22.401 "cfcc4628-d121-4024-b9b1-d14276d72b97" 00:10:22.401 ], 00:10:22.401 "product_name": "Malloc disk", 00:10:22.401 "block_size": 512, 00:10:22.401 "num_blocks": 65536, 00:10:22.401 "uuid": "cfcc4628-d121-4024-b9b1-d14276d72b97", 00:10:22.401 "assigned_rate_limits": { 00:10:22.401 "rw_ios_per_sec": 0, 00:10:22.401 "rw_mbytes_per_sec": 0, 00:10:22.401 "r_mbytes_per_sec": 0, 00:10:22.401 "w_mbytes_per_sec": 0 00:10:22.402 }, 00:10:22.402 "claimed": true, 00:10:22.402 "claim_type": "exclusive_write", 00:10:22.402 "zoned": false, 00:10:22.402 "supported_io_types": { 00:10:22.402 "read": true, 00:10:22.402 "write": true, 00:10:22.402 "unmap": true, 00:10:22.402 "flush": true, 00:10:22.402 "reset": true, 00:10:22.402 "nvme_admin": false, 00:10:22.402 "nvme_io": false, 00:10:22.402 "nvme_io_md": false, 00:10:22.402 "write_zeroes": true, 00:10:22.402 "zcopy": true, 00:10:22.402 "get_zone_info": false, 00:10:22.402 "zone_management": false, 00:10:22.402 "zone_append": false, 00:10:22.402 "compare": false, 00:10:22.402 "compare_and_write": false, 00:10:22.402 "abort": true, 00:10:22.402 "seek_hole": false, 00:10:22.402 "seek_data": false, 00:10:22.402 "copy": true, 00:10:22.402 "nvme_iov_md": false 00:10:22.402 }, 00:10:22.402 "memory_domains": [ 00:10:22.402 { 00:10:22.402 "dma_device_id": "system", 00:10:22.402 "dma_device_type": 1 00:10:22.402 }, 00:10:22.402 { 00:10:22.402 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.402 "dma_device_type": 2 00:10:22.402 } 00:10:22.402 ], 00:10:22.402 "driver_specific": {} 00:10:22.402 } 00:10:22.402 ] 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:22.402 "name": "Existed_Raid", 00:10:22.402 "uuid": "0d373c1f-8a94-4359-a74d-897c99420b35", 00:10:22.402 "strip_size_kb": 64, 00:10:22.402 "state": "online", 00:10:22.402 "raid_level": "concat", 00:10:22.402 "superblock": true, 00:10:22.402 "num_base_bdevs": 4, 00:10:22.402 "num_base_bdevs_discovered": 4, 00:10:22.402 "num_base_bdevs_operational": 4, 00:10:22.402 "base_bdevs_list": [ 00:10:22.402 { 00:10:22.402 "name": "NewBaseBdev", 00:10:22.402 "uuid": "cfcc4628-d121-4024-b9b1-d14276d72b97", 00:10:22.402 "is_configured": true, 00:10:22.402 "data_offset": 2048, 00:10:22.402 "data_size": 63488 00:10:22.402 }, 00:10:22.402 { 00:10:22.402 "name": "BaseBdev2", 00:10:22.402 "uuid": "3e1433a9-d073-4e50-8bef-d72b5541329e", 00:10:22.402 "is_configured": true, 00:10:22.402 "data_offset": 2048, 00:10:22.402 "data_size": 63488 00:10:22.402 }, 00:10:22.402 { 00:10:22.402 "name": "BaseBdev3", 00:10:22.402 "uuid": "dfb40bf8-0670-4bdf-8f4b-83836990c11d", 00:10:22.402 "is_configured": true, 00:10:22.402 "data_offset": 2048, 00:10:22.402 "data_size": 63488 00:10:22.402 }, 00:10:22.402 { 00:10:22.402 "name": "BaseBdev4", 00:10:22.402 "uuid": "bd9be15d-6afe-4951-8b3c-aac2b29119f4", 00:10:22.402 "is_configured": true, 00:10:22.402 "data_offset": 2048, 00:10:22.402 "data_size": 63488 00:10:22.402 } 00:10:22.402 ] 00:10:22.402 }' 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:22.402 19:00:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.972 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:22.972 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:22.972 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:22.972 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:22.972 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:22.972 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:22.972 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:22.972 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:22.972 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.972 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.972 [2024-12-05 19:00:40.236444] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:22.972 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.972 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:22.972 "name": "Existed_Raid", 00:10:22.972 "aliases": [ 00:10:22.972 "0d373c1f-8a94-4359-a74d-897c99420b35" 00:10:22.972 ], 00:10:22.972 "product_name": "Raid Volume", 00:10:22.972 "block_size": 512, 00:10:22.972 "num_blocks": 253952, 00:10:22.972 "uuid": "0d373c1f-8a94-4359-a74d-897c99420b35", 00:10:22.972 "assigned_rate_limits": { 00:10:22.972 "rw_ios_per_sec": 0, 00:10:22.972 "rw_mbytes_per_sec": 0, 00:10:22.972 "r_mbytes_per_sec": 0, 00:10:22.972 "w_mbytes_per_sec": 0 00:10:22.972 }, 00:10:22.972 "claimed": false, 00:10:22.972 "zoned": false, 00:10:22.972 "supported_io_types": { 00:10:22.972 "read": true, 00:10:22.972 "write": true, 00:10:22.972 "unmap": true, 00:10:22.972 "flush": true, 00:10:22.972 "reset": true, 00:10:22.972 "nvme_admin": false, 00:10:22.972 "nvme_io": false, 00:10:22.972 "nvme_io_md": false, 00:10:22.972 "write_zeroes": true, 00:10:22.972 "zcopy": false, 00:10:22.972 "get_zone_info": false, 00:10:22.972 "zone_management": false, 00:10:22.972 "zone_append": false, 00:10:22.972 "compare": false, 00:10:22.972 "compare_and_write": false, 00:10:22.972 "abort": false, 00:10:22.972 "seek_hole": false, 00:10:22.972 "seek_data": false, 00:10:22.972 "copy": false, 00:10:22.972 "nvme_iov_md": false 00:10:22.972 }, 00:10:22.972 "memory_domains": [ 00:10:22.972 { 00:10:22.972 "dma_device_id": "system", 00:10:22.972 "dma_device_type": 1 00:10:22.972 }, 00:10:22.972 { 00:10:22.972 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.972 "dma_device_type": 2 00:10:22.972 }, 00:10:22.972 { 00:10:22.972 "dma_device_id": "system", 00:10:22.972 "dma_device_type": 1 00:10:22.972 }, 00:10:22.972 { 00:10:22.972 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.972 "dma_device_type": 2 00:10:22.972 }, 00:10:22.972 { 00:10:22.972 "dma_device_id": "system", 00:10:22.972 "dma_device_type": 1 00:10:22.972 }, 00:10:22.972 { 00:10:22.972 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.972 "dma_device_type": 2 00:10:22.972 }, 00:10:22.972 { 00:10:22.972 "dma_device_id": "system", 00:10:22.972 "dma_device_type": 1 00:10:22.972 }, 00:10:22.972 { 00:10:22.972 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.972 "dma_device_type": 2 00:10:22.972 } 00:10:22.972 ], 00:10:22.972 "driver_specific": { 00:10:22.972 "raid": { 00:10:22.972 "uuid": "0d373c1f-8a94-4359-a74d-897c99420b35", 00:10:22.972 "strip_size_kb": 64, 00:10:22.972 "state": "online", 00:10:22.972 "raid_level": "concat", 00:10:22.972 "superblock": true, 00:10:22.972 "num_base_bdevs": 4, 00:10:22.972 "num_base_bdevs_discovered": 4, 00:10:22.972 "num_base_bdevs_operational": 4, 00:10:22.972 "base_bdevs_list": [ 00:10:22.972 { 00:10:22.972 "name": "NewBaseBdev", 00:10:22.972 "uuid": "cfcc4628-d121-4024-b9b1-d14276d72b97", 00:10:22.972 "is_configured": true, 00:10:22.972 "data_offset": 2048, 00:10:22.972 "data_size": 63488 00:10:22.972 }, 00:10:22.972 { 00:10:22.972 "name": "BaseBdev2", 00:10:22.972 "uuid": "3e1433a9-d073-4e50-8bef-d72b5541329e", 00:10:22.972 "is_configured": true, 00:10:22.972 "data_offset": 2048, 00:10:22.972 "data_size": 63488 00:10:22.972 }, 00:10:22.972 { 00:10:22.972 "name": "BaseBdev3", 00:10:22.972 "uuid": "dfb40bf8-0670-4bdf-8f4b-83836990c11d", 00:10:22.972 "is_configured": true, 00:10:22.972 "data_offset": 2048, 00:10:22.972 "data_size": 63488 00:10:22.972 }, 00:10:22.972 { 00:10:22.973 "name": "BaseBdev4", 00:10:22.973 "uuid": "bd9be15d-6afe-4951-8b3c-aac2b29119f4", 00:10:22.973 "is_configured": true, 00:10:22.973 "data_offset": 2048, 00:10:22.973 "data_size": 63488 00:10:22.973 } 00:10:22.973 ] 00:10:22.973 } 00:10:22.973 } 00:10:22.973 }' 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:22.973 BaseBdev2 00:10:22.973 BaseBdev3 00:10:22.973 BaseBdev4' 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:22.973 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.233 [2024-12-05 19:00:40.543878] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:23.233 [2024-12-05 19:00:40.543951] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:23.233 [2024-12-05 19:00:40.544035] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:23.233 [2024-12-05 19:00:40.544117] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:23.233 [2024-12-05 19:00:40.544149] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 81897 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 81897 ']' 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 81897 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 81897 00:10:23.233 killing process with pid 81897 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 81897' 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 81897 00:10:23.233 [2024-12-05 19:00:40.584083] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:23.233 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 81897 00:10:23.233 [2024-12-05 19:00:40.623484] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:23.492 19:00:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:23.492 00:10:23.492 real 0m8.637s 00:10:23.492 user 0m14.741s 00:10:23.492 sys 0m1.725s 00:10:23.492 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:23.493 19:00:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:23.493 ************************************ 00:10:23.493 END TEST raid_state_function_test_sb 00:10:23.493 ************************************ 00:10:23.493 19:00:40 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 4 00:10:23.493 19:00:40 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:10:23.493 19:00:40 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:23.493 19:00:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:23.493 ************************************ 00:10:23.493 START TEST raid_superblock_test 00:10:23.493 ************************************ 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test concat 4 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=82486 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 82486 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 82486 ']' 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:23.493 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:23.493 19:00:40 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.493 [2024-12-05 19:00:40.990933] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:10:23.493 [2024-12-05 19:00:40.991147] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82486 ] 00:10:23.751 [2024-12-05 19:00:41.147108] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:23.751 [2024-12-05 19:00:41.171912] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:23.751 [2024-12-05 19:00:41.215193] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:23.751 [2024-12-05 19:00:41.215320] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.319 malloc1 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.319 [2024-12-05 19:00:41.831468] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:24.319 [2024-12-05 19:00:41.831529] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:24.319 [2024-12-05 19:00:41.831549] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:24.319 [2024-12-05 19:00:41.831562] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:24.319 [2024-12-05 19:00:41.833648] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:24.319 [2024-12-05 19:00:41.833700] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:24.319 pt1 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.319 malloc2 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.319 [2024-12-05 19:00:41.860055] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:24.319 [2024-12-05 19:00:41.860170] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:24.319 [2024-12-05 19:00:41.860208] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:24.319 [2024-12-05 19:00:41.860237] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:24.319 [2024-12-05 19:00:41.862295] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:24.319 [2024-12-05 19:00:41.862365] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:24.319 pt2 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.319 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.578 malloc3 00:10:24.578 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.578 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:24.578 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.578 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.578 [2024-12-05 19:00:41.892611] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:24.579 [2024-12-05 19:00:41.892720] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:24.579 [2024-12-05 19:00:41.892772] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:24.579 [2024-12-05 19:00:41.892803] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:24.579 [2024-12-05 19:00:41.894891] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:24.579 [2024-12-05 19:00:41.894961] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:24.579 pt3 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.579 malloc4 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.579 [2024-12-05 19:00:41.947860] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:24.579 [2024-12-05 19:00:41.947959] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:24.579 [2024-12-05 19:00:41.947992] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:24.579 [2024-12-05 19:00:41.948020] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:24.579 [2024-12-05 19:00:41.952366] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:24.579 [2024-12-05 19:00:41.952421] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:24.579 pt4 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.579 [2024-12-05 19:00:41.960658] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:24.579 [2024-12-05 19:00:41.963091] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:24.579 [2024-12-05 19:00:41.963219] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:24.579 [2024-12-05 19:00:41.963300] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:24.579 [2024-12-05 19:00:41.963465] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:24.579 [2024-12-05 19:00:41.963479] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:24.579 [2024-12-05 19:00:41.963759] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:24.579 [2024-12-05 19:00:41.963915] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:24.579 [2024-12-05 19:00:41.963926] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:24.579 [2024-12-05 19:00:41.964047] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.579 19:00:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:24.579 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:24.579 "name": "raid_bdev1", 00:10:24.579 "uuid": "a6dc10d4-54a8-4976-9b35-25a7172c87ef", 00:10:24.579 "strip_size_kb": 64, 00:10:24.579 "state": "online", 00:10:24.579 "raid_level": "concat", 00:10:24.579 "superblock": true, 00:10:24.579 "num_base_bdevs": 4, 00:10:24.579 "num_base_bdevs_discovered": 4, 00:10:24.579 "num_base_bdevs_operational": 4, 00:10:24.579 "base_bdevs_list": [ 00:10:24.579 { 00:10:24.579 "name": "pt1", 00:10:24.579 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:24.579 "is_configured": true, 00:10:24.579 "data_offset": 2048, 00:10:24.579 "data_size": 63488 00:10:24.579 }, 00:10:24.579 { 00:10:24.579 "name": "pt2", 00:10:24.579 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:24.579 "is_configured": true, 00:10:24.579 "data_offset": 2048, 00:10:24.579 "data_size": 63488 00:10:24.579 }, 00:10:24.579 { 00:10:24.579 "name": "pt3", 00:10:24.579 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:24.579 "is_configured": true, 00:10:24.579 "data_offset": 2048, 00:10:24.579 "data_size": 63488 00:10:24.579 }, 00:10:24.579 { 00:10:24.579 "name": "pt4", 00:10:24.579 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:24.579 "is_configured": true, 00:10:24.579 "data_offset": 2048, 00:10:24.579 "data_size": 63488 00:10:24.579 } 00:10:24.579 ] 00:10:24.579 }' 00:10:24.579 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:24.579 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.838 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:24.838 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:24.838 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:24.838 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:24.838 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:24.838 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:24.838 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:24.838 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:24.838 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:24.838 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.838 [2024-12-05 19:00:42.372174] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:24.838 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.097 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:25.097 "name": "raid_bdev1", 00:10:25.097 "aliases": [ 00:10:25.097 "a6dc10d4-54a8-4976-9b35-25a7172c87ef" 00:10:25.097 ], 00:10:25.097 "product_name": "Raid Volume", 00:10:25.097 "block_size": 512, 00:10:25.097 "num_blocks": 253952, 00:10:25.097 "uuid": "a6dc10d4-54a8-4976-9b35-25a7172c87ef", 00:10:25.097 "assigned_rate_limits": { 00:10:25.097 "rw_ios_per_sec": 0, 00:10:25.097 "rw_mbytes_per_sec": 0, 00:10:25.097 "r_mbytes_per_sec": 0, 00:10:25.097 "w_mbytes_per_sec": 0 00:10:25.097 }, 00:10:25.097 "claimed": false, 00:10:25.097 "zoned": false, 00:10:25.097 "supported_io_types": { 00:10:25.097 "read": true, 00:10:25.097 "write": true, 00:10:25.097 "unmap": true, 00:10:25.097 "flush": true, 00:10:25.097 "reset": true, 00:10:25.097 "nvme_admin": false, 00:10:25.097 "nvme_io": false, 00:10:25.097 "nvme_io_md": false, 00:10:25.097 "write_zeroes": true, 00:10:25.097 "zcopy": false, 00:10:25.097 "get_zone_info": false, 00:10:25.097 "zone_management": false, 00:10:25.097 "zone_append": false, 00:10:25.097 "compare": false, 00:10:25.097 "compare_and_write": false, 00:10:25.097 "abort": false, 00:10:25.097 "seek_hole": false, 00:10:25.097 "seek_data": false, 00:10:25.097 "copy": false, 00:10:25.097 "nvme_iov_md": false 00:10:25.097 }, 00:10:25.097 "memory_domains": [ 00:10:25.097 { 00:10:25.097 "dma_device_id": "system", 00:10:25.097 "dma_device_type": 1 00:10:25.097 }, 00:10:25.097 { 00:10:25.097 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.097 "dma_device_type": 2 00:10:25.097 }, 00:10:25.097 { 00:10:25.097 "dma_device_id": "system", 00:10:25.097 "dma_device_type": 1 00:10:25.097 }, 00:10:25.097 { 00:10:25.097 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.097 "dma_device_type": 2 00:10:25.097 }, 00:10:25.097 { 00:10:25.097 "dma_device_id": "system", 00:10:25.097 "dma_device_type": 1 00:10:25.097 }, 00:10:25.097 { 00:10:25.097 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.097 "dma_device_type": 2 00:10:25.097 }, 00:10:25.097 { 00:10:25.097 "dma_device_id": "system", 00:10:25.097 "dma_device_type": 1 00:10:25.097 }, 00:10:25.097 { 00:10:25.097 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.097 "dma_device_type": 2 00:10:25.097 } 00:10:25.097 ], 00:10:25.097 "driver_specific": { 00:10:25.097 "raid": { 00:10:25.097 "uuid": "a6dc10d4-54a8-4976-9b35-25a7172c87ef", 00:10:25.097 "strip_size_kb": 64, 00:10:25.097 "state": "online", 00:10:25.097 "raid_level": "concat", 00:10:25.097 "superblock": true, 00:10:25.097 "num_base_bdevs": 4, 00:10:25.097 "num_base_bdevs_discovered": 4, 00:10:25.097 "num_base_bdevs_operational": 4, 00:10:25.097 "base_bdevs_list": [ 00:10:25.097 { 00:10:25.097 "name": "pt1", 00:10:25.097 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:25.097 "is_configured": true, 00:10:25.097 "data_offset": 2048, 00:10:25.097 "data_size": 63488 00:10:25.097 }, 00:10:25.097 { 00:10:25.097 "name": "pt2", 00:10:25.097 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:25.097 "is_configured": true, 00:10:25.097 "data_offset": 2048, 00:10:25.098 "data_size": 63488 00:10:25.098 }, 00:10:25.098 { 00:10:25.098 "name": "pt3", 00:10:25.098 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:25.098 "is_configured": true, 00:10:25.098 "data_offset": 2048, 00:10:25.098 "data_size": 63488 00:10:25.098 }, 00:10:25.098 { 00:10:25.098 "name": "pt4", 00:10:25.098 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:25.098 "is_configured": true, 00:10:25.098 "data_offset": 2048, 00:10:25.098 "data_size": 63488 00:10:25.098 } 00:10:25.098 ] 00:10:25.098 } 00:10:25.098 } 00:10:25.098 }' 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:25.098 pt2 00:10:25.098 pt3 00:10:25.098 pt4' 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.098 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.358 [2024-12-05 19:00:42.691572] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=a6dc10d4-54a8-4976-9b35-25a7172c87ef 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z a6dc10d4-54a8-4976-9b35-25a7172c87ef ']' 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.358 [2024-12-05 19:00:42.723247] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:25.358 [2024-12-05 19:00:42.723276] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:25.358 [2024-12-05 19:00:42.723344] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:25.358 [2024-12-05 19:00:42.723413] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:25.358 [2024-12-05 19:00:42.723423] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.358 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.358 [2024-12-05 19:00:42.863030] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:25.358 [2024-12-05 19:00:42.864844] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:25.358 [2024-12-05 19:00:42.864889] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:25.358 [2024-12-05 19:00:42.864917] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:10:25.358 [2024-12-05 19:00:42.864960] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:25.358 [2024-12-05 19:00:42.865000] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:25.358 [2024-12-05 19:00:42.865018] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:25.358 [2024-12-05 19:00:42.865033] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:10:25.358 [2024-12-05 19:00:42.865046] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:25.358 [2024-12-05 19:00:42.865063] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:10:25.358 request: 00:10:25.358 { 00:10:25.358 "name": "raid_bdev1", 00:10:25.358 "raid_level": "concat", 00:10:25.358 "base_bdevs": [ 00:10:25.358 "malloc1", 00:10:25.358 "malloc2", 00:10:25.358 "malloc3", 00:10:25.358 "malloc4" 00:10:25.358 ], 00:10:25.358 "strip_size_kb": 64, 00:10:25.358 "superblock": false, 00:10:25.359 "method": "bdev_raid_create", 00:10:25.359 "req_id": 1 00:10:25.359 } 00:10:25.359 Got JSON-RPC error response 00:10:25.359 response: 00:10:25.359 { 00:10:25.359 "code": -17, 00:10:25.359 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:25.359 } 00:10:25.359 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:10:25.359 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:10:25.359 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:10:25.359 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:10:25.359 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:10:25.359 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.359 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.359 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.359 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:25.359 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.643 [2024-12-05 19:00:42.930883] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:25.643 [2024-12-05 19:00:42.930975] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:25.643 [2024-12-05 19:00:42.931016] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:25.643 [2024-12-05 19:00:42.931043] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:25.643 [2024-12-05 19:00:42.933106] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:25.643 [2024-12-05 19:00:42.933176] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:25.643 [2024-12-05 19:00:42.933257] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:25.643 [2024-12-05 19:00:42.933318] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:25.643 pt1 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.643 "name": "raid_bdev1", 00:10:25.643 "uuid": "a6dc10d4-54a8-4976-9b35-25a7172c87ef", 00:10:25.643 "strip_size_kb": 64, 00:10:25.643 "state": "configuring", 00:10:25.643 "raid_level": "concat", 00:10:25.643 "superblock": true, 00:10:25.643 "num_base_bdevs": 4, 00:10:25.643 "num_base_bdevs_discovered": 1, 00:10:25.643 "num_base_bdevs_operational": 4, 00:10:25.643 "base_bdevs_list": [ 00:10:25.643 { 00:10:25.643 "name": "pt1", 00:10:25.643 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:25.643 "is_configured": true, 00:10:25.643 "data_offset": 2048, 00:10:25.643 "data_size": 63488 00:10:25.643 }, 00:10:25.643 { 00:10:25.643 "name": null, 00:10:25.643 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:25.643 "is_configured": false, 00:10:25.643 "data_offset": 2048, 00:10:25.643 "data_size": 63488 00:10:25.643 }, 00:10:25.643 { 00:10:25.643 "name": null, 00:10:25.643 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:25.643 "is_configured": false, 00:10:25.643 "data_offset": 2048, 00:10:25.643 "data_size": 63488 00:10:25.643 }, 00:10:25.643 { 00:10:25.643 "name": null, 00:10:25.643 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:25.643 "is_configured": false, 00:10:25.643 "data_offset": 2048, 00:10:25.643 "data_size": 63488 00:10:25.643 } 00:10:25.643 ] 00:10:25.643 }' 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.643 19:00:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.903 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:25.903 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:25.903 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.903 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.903 [2024-12-05 19:00:43.326208] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:25.903 [2024-12-05 19:00:43.326258] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:25.903 [2024-12-05 19:00:43.326276] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:10:25.903 [2024-12-05 19:00:43.326284] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:25.903 [2024-12-05 19:00:43.326619] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:25.903 [2024-12-05 19:00:43.326635] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:25.903 [2024-12-05 19:00:43.326716] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:25.903 [2024-12-05 19:00:43.326736] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:25.903 pt2 00:10:25.903 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.903 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.904 [2024-12-05 19:00:43.338216] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.904 "name": "raid_bdev1", 00:10:25.904 "uuid": "a6dc10d4-54a8-4976-9b35-25a7172c87ef", 00:10:25.904 "strip_size_kb": 64, 00:10:25.904 "state": "configuring", 00:10:25.904 "raid_level": "concat", 00:10:25.904 "superblock": true, 00:10:25.904 "num_base_bdevs": 4, 00:10:25.904 "num_base_bdevs_discovered": 1, 00:10:25.904 "num_base_bdevs_operational": 4, 00:10:25.904 "base_bdevs_list": [ 00:10:25.904 { 00:10:25.904 "name": "pt1", 00:10:25.904 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:25.904 "is_configured": true, 00:10:25.904 "data_offset": 2048, 00:10:25.904 "data_size": 63488 00:10:25.904 }, 00:10:25.904 { 00:10:25.904 "name": null, 00:10:25.904 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:25.904 "is_configured": false, 00:10:25.904 "data_offset": 0, 00:10:25.904 "data_size": 63488 00:10:25.904 }, 00:10:25.904 { 00:10:25.904 "name": null, 00:10:25.904 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:25.904 "is_configured": false, 00:10:25.904 "data_offset": 2048, 00:10:25.904 "data_size": 63488 00:10:25.904 }, 00:10:25.904 { 00:10:25.904 "name": null, 00:10:25.904 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:25.904 "is_configured": false, 00:10:25.904 "data_offset": 2048, 00:10:25.904 "data_size": 63488 00:10:25.904 } 00:10:25.904 ] 00:10:25.904 }' 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.904 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.163 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:26.163 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:26.163 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:26.163 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.163 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.163 [2024-12-05 19:00:43.717554] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:26.163 [2024-12-05 19:00:43.717659] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:26.163 [2024-12-05 19:00:43.717701] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:10:26.163 [2024-12-05 19:00:43.717740] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:26.163 [2024-12-05 19:00:43.718136] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:26.163 [2024-12-05 19:00:43.718202] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:26.163 [2024-12-05 19:00:43.718297] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:26.163 [2024-12-05 19:00:43.718356] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:26.423 pt2 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.423 [2024-12-05 19:00:43.729498] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:26.423 [2024-12-05 19:00:43.729580] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:26.423 [2024-12-05 19:00:43.729625] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:26.423 [2024-12-05 19:00:43.729652] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:26.423 [2024-12-05 19:00:43.730003] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:26.423 [2024-12-05 19:00:43.730060] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:26.423 [2024-12-05 19:00:43.730138] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:26.423 [2024-12-05 19:00:43.730184] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:26.423 pt3 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.423 [2024-12-05 19:00:43.741482] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:26.423 [2024-12-05 19:00:43.741531] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:26.423 [2024-12-05 19:00:43.741559] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:10:26.423 [2024-12-05 19:00:43.741568] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:26.423 [2024-12-05 19:00:43.741859] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:26.423 [2024-12-05 19:00:43.741877] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:26.423 [2024-12-05 19:00:43.741925] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:26.423 [2024-12-05 19:00:43.741943] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:26.423 [2024-12-05 19:00:43.742030] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:26.423 [2024-12-05 19:00:43.742040] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:26.423 [2024-12-05 19:00:43.742248] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:10:26.423 [2024-12-05 19:00:43.742358] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:26.423 [2024-12-05 19:00:43.742365] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:10:26.423 [2024-12-05 19:00:43.742453] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:26.423 pt4 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:26.423 "name": "raid_bdev1", 00:10:26.423 "uuid": "a6dc10d4-54a8-4976-9b35-25a7172c87ef", 00:10:26.423 "strip_size_kb": 64, 00:10:26.423 "state": "online", 00:10:26.423 "raid_level": "concat", 00:10:26.423 "superblock": true, 00:10:26.423 "num_base_bdevs": 4, 00:10:26.423 "num_base_bdevs_discovered": 4, 00:10:26.423 "num_base_bdevs_operational": 4, 00:10:26.423 "base_bdevs_list": [ 00:10:26.423 { 00:10:26.423 "name": "pt1", 00:10:26.423 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:26.423 "is_configured": true, 00:10:26.423 "data_offset": 2048, 00:10:26.423 "data_size": 63488 00:10:26.423 }, 00:10:26.423 { 00:10:26.423 "name": "pt2", 00:10:26.423 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:26.423 "is_configured": true, 00:10:26.423 "data_offset": 2048, 00:10:26.423 "data_size": 63488 00:10:26.423 }, 00:10:26.423 { 00:10:26.423 "name": "pt3", 00:10:26.423 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:26.423 "is_configured": true, 00:10:26.423 "data_offset": 2048, 00:10:26.423 "data_size": 63488 00:10:26.423 }, 00:10:26.423 { 00:10:26.423 "name": "pt4", 00:10:26.423 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:26.423 "is_configured": true, 00:10:26.423 "data_offset": 2048, 00:10:26.423 "data_size": 63488 00:10:26.423 } 00:10:26.423 ] 00:10:26.423 }' 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:26.423 19:00:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.683 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:26.683 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:26.683 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:26.683 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:26.683 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:26.683 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:26.683 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:26.683 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:26.683 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.683 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.683 [2024-12-05 19:00:44.129122] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:26.683 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.683 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:26.683 "name": "raid_bdev1", 00:10:26.683 "aliases": [ 00:10:26.683 "a6dc10d4-54a8-4976-9b35-25a7172c87ef" 00:10:26.683 ], 00:10:26.683 "product_name": "Raid Volume", 00:10:26.683 "block_size": 512, 00:10:26.683 "num_blocks": 253952, 00:10:26.683 "uuid": "a6dc10d4-54a8-4976-9b35-25a7172c87ef", 00:10:26.683 "assigned_rate_limits": { 00:10:26.683 "rw_ios_per_sec": 0, 00:10:26.683 "rw_mbytes_per_sec": 0, 00:10:26.683 "r_mbytes_per_sec": 0, 00:10:26.683 "w_mbytes_per_sec": 0 00:10:26.683 }, 00:10:26.683 "claimed": false, 00:10:26.683 "zoned": false, 00:10:26.683 "supported_io_types": { 00:10:26.684 "read": true, 00:10:26.684 "write": true, 00:10:26.684 "unmap": true, 00:10:26.684 "flush": true, 00:10:26.684 "reset": true, 00:10:26.684 "nvme_admin": false, 00:10:26.684 "nvme_io": false, 00:10:26.684 "nvme_io_md": false, 00:10:26.684 "write_zeroes": true, 00:10:26.684 "zcopy": false, 00:10:26.684 "get_zone_info": false, 00:10:26.684 "zone_management": false, 00:10:26.684 "zone_append": false, 00:10:26.684 "compare": false, 00:10:26.684 "compare_and_write": false, 00:10:26.684 "abort": false, 00:10:26.684 "seek_hole": false, 00:10:26.684 "seek_data": false, 00:10:26.684 "copy": false, 00:10:26.684 "nvme_iov_md": false 00:10:26.684 }, 00:10:26.684 "memory_domains": [ 00:10:26.684 { 00:10:26.684 "dma_device_id": "system", 00:10:26.684 "dma_device_type": 1 00:10:26.684 }, 00:10:26.684 { 00:10:26.684 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.684 "dma_device_type": 2 00:10:26.684 }, 00:10:26.684 { 00:10:26.684 "dma_device_id": "system", 00:10:26.684 "dma_device_type": 1 00:10:26.684 }, 00:10:26.684 { 00:10:26.684 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.684 "dma_device_type": 2 00:10:26.684 }, 00:10:26.684 { 00:10:26.684 "dma_device_id": "system", 00:10:26.684 "dma_device_type": 1 00:10:26.684 }, 00:10:26.684 { 00:10:26.684 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.684 "dma_device_type": 2 00:10:26.684 }, 00:10:26.684 { 00:10:26.684 "dma_device_id": "system", 00:10:26.684 "dma_device_type": 1 00:10:26.684 }, 00:10:26.684 { 00:10:26.684 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:26.684 "dma_device_type": 2 00:10:26.684 } 00:10:26.684 ], 00:10:26.684 "driver_specific": { 00:10:26.684 "raid": { 00:10:26.684 "uuid": "a6dc10d4-54a8-4976-9b35-25a7172c87ef", 00:10:26.684 "strip_size_kb": 64, 00:10:26.684 "state": "online", 00:10:26.684 "raid_level": "concat", 00:10:26.684 "superblock": true, 00:10:26.684 "num_base_bdevs": 4, 00:10:26.684 "num_base_bdevs_discovered": 4, 00:10:26.684 "num_base_bdevs_operational": 4, 00:10:26.684 "base_bdevs_list": [ 00:10:26.684 { 00:10:26.684 "name": "pt1", 00:10:26.684 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:26.684 "is_configured": true, 00:10:26.684 "data_offset": 2048, 00:10:26.684 "data_size": 63488 00:10:26.684 }, 00:10:26.684 { 00:10:26.684 "name": "pt2", 00:10:26.684 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:26.684 "is_configured": true, 00:10:26.684 "data_offset": 2048, 00:10:26.684 "data_size": 63488 00:10:26.684 }, 00:10:26.684 { 00:10:26.684 "name": "pt3", 00:10:26.684 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:26.684 "is_configured": true, 00:10:26.684 "data_offset": 2048, 00:10:26.684 "data_size": 63488 00:10:26.684 }, 00:10:26.684 { 00:10:26.684 "name": "pt4", 00:10:26.684 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:26.684 "is_configured": true, 00:10:26.684 "data_offset": 2048, 00:10:26.684 "data_size": 63488 00:10:26.684 } 00:10:26.684 ] 00:10:26.684 } 00:10:26.684 } 00:10:26.684 }' 00:10:26.684 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:26.684 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:26.684 pt2 00:10:26.684 pt3 00:10:26.684 pt4' 00:10:26.684 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:26.684 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:26.684 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:26.684 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:26.684 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:26.684 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.684 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.944 [2024-12-05 19:00:44.408684] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' a6dc10d4-54a8-4976-9b35-25a7172c87ef '!=' a6dc10d4-54a8-4976-9b35-25a7172c87ef ']' 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 82486 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 82486 ']' 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 82486 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82486 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:26.944 killing process with pid 82486 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82486' 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 82486 00:10:26.944 [2024-12-05 19:00:44.479843] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:26.944 [2024-12-05 19:00:44.479918] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:26.944 [2024-12-05 19:00:44.479981] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:26.944 [2024-12-05 19:00:44.479992] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:10:26.944 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 82486 00:10:27.203 [2024-12-05 19:00:44.522412] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:27.203 19:00:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:27.203 00:10:27.203 real 0m3.831s 00:10:27.203 user 0m5.952s 00:10:27.203 sys 0m0.842s 00:10:27.203 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:27.203 19:00:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.203 ************************************ 00:10:27.203 END TEST raid_superblock_test 00:10:27.203 ************************************ 00:10:27.463 19:00:44 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 4 read 00:10:27.463 19:00:44 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:27.463 19:00:44 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:27.463 19:00:44 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:27.463 ************************************ 00:10:27.463 START TEST raid_read_error_test 00:10:27.463 ************************************ 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 4 read 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.2ZncENLMAB 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=82710 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 82710 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 82710 ']' 00:10:27.463 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:27.463 19:00:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:27.464 19:00:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:27.464 19:00:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:27.464 19:00:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.464 [2024-12-05 19:00:44.902722] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:10:27.464 [2024-12-05 19:00:44.902866] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82710 ] 00:10:27.723 [2024-12-05 19:00:45.056005] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:27.724 [2024-12-05 19:00:45.080869] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:27.724 [2024-12-05 19:00:45.124046] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:27.724 [2024-12-05 19:00:45.124079] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.293 BaseBdev1_malloc 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.293 true 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.293 [2024-12-05 19:00:45.756028] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:28.293 [2024-12-05 19:00:45.756082] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:28.293 [2024-12-05 19:00:45.756109] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:28.293 [2024-12-05 19:00:45.756118] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:28.293 [2024-12-05 19:00:45.758271] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:28.293 [2024-12-05 19:00:45.758312] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:28.293 BaseBdev1 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.293 BaseBdev2_malloc 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.293 true 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.293 [2024-12-05 19:00:45.796855] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:28.293 [2024-12-05 19:00:45.796900] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:28.293 [2024-12-05 19:00:45.796918] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:28.293 [2024-12-05 19:00:45.796935] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:28.293 [2024-12-05 19:00:45.799029] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:28.293 [2024-12-05 19:00:45.799126] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:28.293 BaseBdev2 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.293 BaseBdev3_malloc 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.293 true 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.293 [2024-12-05 19:00:45.837585] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:28.293 [2024-12-05 19:00:45.837633] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:28.293 [2024-12-05 19:00:45.837668] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:28.293 [2024-12-05 19:00:45.837676] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:28.293 [2024-12-05 19:00:45.839740] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:28.293 [2024-12-05 19:00:45.839772] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:28.293 BaseBdev3 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.293 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.553 BaseBdev4_malloc 00:10:28.553 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.553 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:28.553 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.553 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.553 true 00:10:28.553 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.553 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:28.553 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.553 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.553 [2024-12-05 19:00:45.897815] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:28.553 [2024-12-05 19:00:45.897882] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:28.553 [2024-12-05 19:00:45.897914] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:28.553 [2024-12-05 19:00:45.897927] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:28.553 [2024-12-05 19:00:45.900710] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:28.553 [2024-12-05 19:00:45.900747] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:28.553 BaseBdev4 00:10:28.553 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.553 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:28.553 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.553 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.554 [2024-12-05 19:00:45.909801] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:28.554 [2024-12-05 19:00:45.911602] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:28.554 [2024-12-05 19:00:45.911682] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:28.554 [2024-12-05 19:00:45.911811] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:28.554 [2024-12-05 19:00:45.912042] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:28.554 [2024-12-05 19:00:45.912093] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:28.554 [2024-12-05 19:00:45.912350] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:10:28.554 [2024-12-05 19:00:45.912497] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:28.554 [2024-12-05 19:00:45.912517] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:28.554 [2024-12-05 19:00:45.912652] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:28.554 "name": "raid_bdev1", 00:10:28.554 "uuid": "1d536501-9d36-420f-9502-9c618af4897c", 00:10:28.554 "strip_size_kb": 64, 00:10:28.554 "state": "online", 00:10:28.554 "raid_level": "concat", 00:10:28.554 "superblock": true, 00:10:28.554 "num_base_bdevs": 4, 00:10:28.554 "num_base_bdevs_discovered": 4, 00:10:28.554 "num_base_bdevs_operational": 4, 00:10:28.554 "base_bdevs_list": [ 00:10:28.554 { 00:10:28.554 "name": "BaseBdev1", 00:10:28.554 "uuid": "77f83870-e825-5c1d-9e21-039f7c8c270a", 00:10:28.554 "is_configured": true, 00:10:28.554 "data_offset": 2048, 00:10:28.554 "data_size": 63488 00:10:28.554 }, 00:10:28.554 { 00:10:28.554 "name": "BaseBdev2", 00:10:28.554 "uuid": "3e12757d-0d1d-506d-a342-a025433c405f", 00:10:28.554 "is_configured": true, 00:10:28.554 "data_offset": 2048, 00:10:28.554 "data_size": 63488 00:10:28.554 }, 00:10:28.554 { 00:10:28.554 "name": "BaseBdev3", 00:10:28.554 "uuid": "951f71e6-febf-512a-bc06-b25ba4535357", 00:10:28.554 "is_configured": true, 00:10:28.554 "data_offset": 2048, 00:10:28.554 "data_size": 63488 00:10:28.554 }, 00:10:28.554 { 00:10:28.554 "name": "BaseBdev4", 00:10:28.554 "uuid": "6bdddbe7-d06c-518c-a2f0-edfc28f3b4ca", 00:10:28.554 "is_configured": true, 00:10:28.554 "data_offset": 2048, 00:10:28.554 "data_size": 63488 00:10:28.554 } 00:10:28.554 ] 00:10:28.554 }' 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:28.554 19:00:45 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.814 19:00:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:28.814 19:00:46 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:28.814 [2024-12-05 19:00:46.337365] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.754 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.043 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:30.043 "name": "raid_bdev1", 00:10:30.043 "uuid": "1d536501-9d36-420f-9502-9c618af4897c", 00:10:30.043 "strip_size_kb": 64, 00:10:30.043 "state": "online", 00:10:30.043 "raid_level": "concat", 00:10:30.043 "superblock": true, 00:10:30.043 "num_base_bdevs": 4, 00:10:30.043 "num_base_bdevs_discovered": 4, 00:10:30.043 "num_base_bdevs_operational": 4, 00:10:30.043 "base_bdevs_list": [ 00:10:30.043 { 00:10:30.043 "name": "BaseBdev1", 00:10:30.043 "uuid": "77f83870-e825-5c1d-9e21-039f7c8c270a", 00:10:30.043 "is_configured": true, 00:10:30.043 "data_offset": 2048, 00:10:30.043 "data_size": 63488 00:10:30.043 }, 00:10:30.043 { 00:10:30.043 "name": "BaseBdev2", 00:10:30.043 "uuid": "3e12757d-0d1d-506d-a342-a025433c405f", 00:10:30.043 "is_configured": true, 00:10:30.043 "data_offset": 2048, 00:10:30.043 "data_size": 63488 00:10:30.043 }, 00:10:30.043 { 00:10:30.043 "name": "BaseBdev3", 00:10:30.043 "uuid": "951f71e6-febf-512a-bc06-b25ba4535357", 00:10:30.043 "is_configured": true, 00:10:30.043 "data_offset": 2048, 00:10:30.043 "data_size": 63488 00:10:30.043 }, 00:10:30.044 { 00:10:30.044 "name": "BaseBdev4", 00:10:30.044 "uuid": "6bdddbe7-d06c-518c-a2f0-edfc28f3b4ca", 00:10:30.044 "is_configured": true, 00:10:30.044 "data_offset": 2048, 00:10:30.044 "data_size": 63488 00:10:30.044 } 00:10:30.044 ] 00:10:30.044 }' 00:10:30.044 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:30.044 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.304 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:30.304 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:30.304 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.304 [2024-12-05 19:00:47.709299] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:30.304 [2024-12-05 19:00:47.709403] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:30.304 [2024-12-05 19:00:47.711917] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:30.304 [2024-12-05 19:00:47.711968] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:30.304 [2024-12-05 19:00:47.712011] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:30.304 [2024-12-05 19:00:47.712021] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:30.304 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:30.304 { 00:10:30.304 "results": [ 00:10:30.304 { 00:10:30.304 "job": "raid_bdev1", 00:10:30.304 "core_mask": "0x1", 00:10:30.304 "workload": "randrw", 00:10:30.304 "percentage": 50, 00:10:30.304 "status": "finished", 00:10:30.304 "queue_depth": 1, 00:10:30.304 "io_size": 131072, 00:10:30.304 "runtime": 1.372876, 00:10:30.304 "iops": 16871.1522380754, 00:10:30.304 "mibps": 2108.894029759425, 00:10:30.304 "io_failed": 1, 00:10:30.304 "io_timeout": 0, 00:10:30.304 "avg_latency_us": 81.9709994500716, 00:10:30.304 "min_latency_us": 24.929257641921396, 00:10:30.304 "max_latency_us": 1402.2986899563318 00:10:30.304 } 00:10:30.304 ], 00:10:30.304 "core_count": 1 00:10:30.304 } 00:10:30.304 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 82710 00:10:30.304 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 82710 ']' 00:10:30.304 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 82710 00:10:30.304 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:10:30.304 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:30.304 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82710 00:10:30.304 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:30.304 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:30.304 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82710' 00:10:30.304 killing process with pid 82710 00:10:30.305 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 82710 00:10:30.305 [2024-12-05 19:00:47.745455] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:30.305 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 82710 00:10:30.305 [2024-12-05 19:00:47.779423] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:30.565 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.2ZncENLMAB 00:10:30.565 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:30.565 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:30.565 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:10:30.565 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:30.565 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:30.565 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:30.565 19:00:47 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:10:30.565 00:10:30.565 real 0m3.195s 00:10:30.565 user 0m3.966s 00:10:30.565 sys 0m0.491s 00:10:30.565 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:30.565 19:00:47 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.565 ************************************ 00:10:30.565 END TEST raid_read_error_test 00:10:30.565 ************************************ 00:10:30.565 19:00:48 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 4 write 00:10:30.565 19:00:48 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:30.565 19:00:48 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:30.565 19:00:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:30.565 ************************************ 00:10:30.565 START TEST raid_write_error_test 00:10:30.565 ************************************ 00:10:30.565 19:00:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test concat 4 write 00:10:30.565 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:30.565 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:30.565 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:30.565 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.sJnY8Do6fm 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=82822 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 82822 00:10:30.566 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 82822 ']' 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:30.566 19:00:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.826 [2024-12-05 19:00:48.168267] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:10:30.826 [2024-12-05 19:00:48.168414] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82822 ] 00:10:30.826 [2024-12-05 19:00:48.299919] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:30.826 [2024-12-05 19:00:48.324655] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:30.826 [2024-12-05 19:00:48.367602] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:30.826 [2024-12-05 19:00:48.367644] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:31.782 19:00:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:31.782 19:00:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:10:31.782 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:31.782 19:00:48 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:31.782 19:00:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.782 19:00:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.782 BaseBdev1_malloc 00:10:31.782 19:00:48 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.782 true 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.782 [2024-12-05 19:00:49.019486] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:31.782 [2024-12-05 19:00:49.019536] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:31.782 [2024-12-05 19:00:49.019563] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:31.782 [2024-12-05 19:00:49.019578] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:31.782 [2024-12-05 19:00:49.021785] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:31.782 [2024-12-05 19:00:49.021818] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:31.782 BaseBdev1 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.782 BaseBdev2_malloc 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.782 true 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.782 [2024-12-05 19:00:49.060319] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:31.782 [2024-12-05 19:00:49.060377] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:31.782 [2024-12-05 19:00:49.060396] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:31.782 [2024-12-05 19:00:49.060412] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:31.782 [2024-12-05 19:00:49.062504] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:31.782 [2024-12-05 19:00:49.062573] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:31.782 BaseBdev2 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.782 BaseBdev3_malloc 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.782 true 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.782 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.782 [2024-12-05 19:00:49.100855] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:31.783 [2024-12-05 19:00:49.100945] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:31.783 [2024-12-05 19:00:49.100984] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:31.783 [2024-12-05 19:00:49.100992] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:31.783 [2024-12-05 19:00:49.102995] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:31.783 [2024-12-05 19:00:49.103030] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:31.783 BaseBdev3 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.783 BaseBdev4_malloc 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.783 true 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.783 [2024-12-05 19:00:49.149525] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:31.783 [2024-12-05 19:00:49.149572] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:31.783 [2024-12-05 19:00:49.149593] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:31.783 [2024-12-05 19:00:49.149601] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:31.783 [2024-12-05 19:00:49.151597] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:31.783 [2024-12-05 19:00:49.151633] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:31.783 BaseBdev4 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.783 [2024-12-05 19:00:49.161543] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:31.783 [2024-12-05 19:00:49.163333] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:31.783 [2024-12-05 19:00:49.163419] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:31.783 [2024-12-05 19:00:49.163476] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:31.783 [2024-12-05 19:00:49.163658] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:31.783 [2024-12-05 19:00:49.163681] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:31.783 [2024-12-05 19:00:49.163937] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:10:31.783 [2024-12-05 19:00:49.164072] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:31.783 [2024-12-05 19:00:49.164085] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:31.783 [2024-12-05 19:00:49.164199] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:31.783 "name": "raid_bdev1", 00:10:31.783 "uuid": "2e7a93d9-e38d-4e0a-8823-297d88f22f31", 00:10:31.783 "strip_size_kb": 64, 00:10:31.783 "state": "online", 00:10:31.783 "raid_level": "concat", 00:10:31.783 "superblock": true, 00:10:31.783 "num_base_bdevs": 4, 00:10:31.783 "num_base_bdevs_discovered": 4, 00:10:31.783 "num_base_bdevs_operational": 4, 00:10:31.783 "base_bdevs_list": [ 00:10:31.783 { 00:10:31.783 "name": "BaseBdev1", 00:10:31.783 "uuid": "e2a98386-d692-59f5-88b4-72f4ab29cf52", 00:10:31.783 "is_configured": true, 00:10:31.783 "data_offset": 2048, 00:10:31.783 "data_size": 63488 00:10:31.783 }, 00:10:31.783 { 00:10:31.783 "name": "BaseBdev2", 00:10:31.783 "uuid": "333d4530-dfe2-59fc-846c-287a1e67cae3", 00:10:31.783 "is_configured": true, 00:10:31.783 "data_offset": 2048, 00:10:31.783 "data_size": 63488 00:10:31.783 }, 00:10:31.783 { 00:10:31.783 "name": "BaseBdev3", 00:10:31.783 "uuid": "f2ee4e35-ae5e-5a15-8175-94a687fd89bc", 00:10:31.783 "is_configured": true, 00:10:31.783 "data_offset": 2048, 00:10:31.783 "data_size": 63488 00:10:31.783 }, 00:10:31.783 { 00:10:31.783 "name": "BaseBdev4", 00:10:31.783 "uuid": "0194c07e-1130-508a-a8f4-e7ce73b49c9a", 00:10:31.783 "is_configured": true, 00:10:31.783 "data_offset": 2048, 00:10:31.783 "data_size": 63488 00:10:31.783 } 00:10:31.783 ] 00:10:31.783 }' 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:31.783 19:00:49 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.044 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:32.044 19:00:49 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:32.304 [2024-12-05 19:00:49.681023] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.244 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:33.244 "name": "raid_bdev1", 00:10:33.244 "uuid": "2e7a93d9-e38d-4e0a-8823-297d88f22f31", 00:10:33.244 "strip_size_kb": 64, 00:10:33.244 "state": "online", 00:10:33.244 "raid_level": "concat", 00:10:33.244 "superblock": true, 00:10:33.244 "num_base_bdevs": 4, 00:10:33.244 "num_base_bdevs_discovered": 4, 00:10:33.244 "num_base_bdevs_operational": 4, 00:10:33.244 "base_bdevs_list": [ 00:10:33.244 { 00:10:33.244 "name": "BaseBdev1", 00:10:33.245 "uuid": "e2a98386-d692-59f5-88b4-72f4ab29cf52", 00:10:33.245 "is_configured": true, 00:10:33.245 "data_offset": 2048, 00:10:33.245 "data_size": 63488 00:10:33.245 }, 00:10:33.245 { 00:10:33.245 "name": "BaseBdev2", 00:10:33.245 "uuid": "333d4530-dfe2-59fc-846c-287a1e67cae3", 00:10:33.245 "is_configured": true, 00:10:33.245 "data_offset": 2048, 00:10:33.245 "data_size": 63488 00:10:33.245 }, 00:10:33.245 { 00:10:33.245 "name": "BaseBdev3", 00:10:33.245 "uuid": "f2ee4e35-ae5e-5a15-8175-94a687fd89bc", 00:10:33.245 "is_configured": true, 00:10:33.245 "data_offset": 2048, 00:10:33.245 "data_size": 63488 00:10:33.245 }, 00:10:33.245 { 00:10:33.245 "name": "BaseBdev4", 00:10:33.245 "uuid": "0194c07e-1130-508a-a8f4-e7ce73b49c9a", 00:10:33.245 "is_configured": true, 00:10:33.245 "data_offset": 2048, 00:10:33.245 "data_size": 63488 00:10:33.245 } 00:10:33.245 ] 00:10:33.245 }' 00:10:33.245 19:00:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:33.245 19:00:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.505 19:00:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:33.505 19:00:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:33.505 19:00:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.505 [2024-12-05 19:00:51.028492] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:33.505 [2024-12-05 19:00:51.028589] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:33.505 [2024-12-05 19:00:51.031193] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:33.505 [2024-12-05 19:00:51.031299] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:33.505 [2024-12-05 19:00:51.031365] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:33.505 [2024-12-05 19:00:51.031409] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:33.505 { 00:10:33.505 "results": [ 00:10:33.505 { 00:10:33.505 "job": "raid_bdev1", 00:10:33.505 "core_mask": "0x1", 00:10:33.505 "workload": "randrw", 00:10:33.505 "percentage": 50, 00:10:33.505 "status": "finished", 00:10:33.505 "queue_depth": 1, 00:10:33.505 "io_size": 131072, 00:10:33.505 "runtime": 1.34842, 00:10:33.505 "iops": 16893.10452232984, 00:10:33.505 "mibps": 2111.63806529123, 00:10:33.505 "io_failed": 1, 00:10:33.505 "io_timeout": 0, 00:10:33.505 "avg_latency_us": 81.83751164547158, 00:10:33.505 "min_latency_us": 24.929257641921396, 00:10:33.505 "max_latency_us": 1316.4436681222708 00:10:33.505 } 00:10:33.505 ], 00:10:33.505 "core_count": 1 00:10:33.505 } 00:10:33.505 19:00:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:33.505 19:00:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 82822 00:10:33.505 19:00:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 82822 ']' 00:10:33.505 19:00:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 82822 00:10:33.505 19:00:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:10:33.505 19:00:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:33.505 19:00:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82822 00:10:33.764 19:00:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:33.764 19:00:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:33.764 19:00:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82822' 00:10:33.764 killing process with pid 82822 00:10:33.764 19:00:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 82822 00:10:33.764 [2024-12-05 19:00:51.073343] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:33.764 19:00:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 82822 00:10:33.764 [2024-12-05 19:00:51.107431] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:33.764 19:00:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:33.764 19:00:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.sJnY8Do6fm 00:10:33.764 19:00:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:33.764 19:00:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:10:33.764 19:00:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:33.764 ************************************ 00:10:33.764 END TEST raid_write_error_test 00:10:33.764 ************************************ 00:10:33.764 19:00:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:33.764 19:00:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:33.764 19:00:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:10:33.764 00:10:33.764 real 0m3.250s 00:10:33.764 user 0m4.081s 00:10:33.764 sys 0m0.531s 00:10:33.764 19:00:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:33.764 19:00:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.023 19:00:51 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:10:34.023 19:00:51 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 4 false 00:10:34.023 19:00:51 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:34.023 19:00:51 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:34.023 19:00:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:34.023 ************************************ 00:10:34.023 START TEST raid_state_function_test 00:10:34.023 ************************************ 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 4 false 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:10:34.023 Process raid pid: 82931 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=82931 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 82931' 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 82931 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 82931 ']' 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:34.023 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:34.023 19:00:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.023 [2024-12-05 19:00:51.482366] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:10:34.023 [2024-12-05 19:00:51.482508] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:34.283 [2024-12-05 19:00:51.636815] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:34.283 [2024-12-05 19:00:51.661563] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:34.283 [2024-12-05 19:00:51.704969] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:34.283 [2024-12-05 19:00:51.705007] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.853 [2024-12-05 19:00:52.304505] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:34.853 [2024-12-05 19:00:52.304616] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:34.853 [2024-12-05 19:00:52.304647] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:34.853 [2024-12-05 19:00:52.304683] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:34.853 [2024-12-05 19:00:52.304704] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:34.853 [2024-12-05 19:00:52.304726] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:34.853 [2024-12-05 19:00:52.304767] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:34.853 [2024-12-05 19:00:52.304944] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:34.853 "name": "Existed_Raid", 00:10:34.853 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:34.853 "strip_size_kb": 0, 00:10:34.853 "state": "configuring", 00:10:34.853 "raid_level": "raid1", 00:10:34.853 "superblock": false, 00:10:34.853 "num_base_bdevs": 4, 00:10:34.853 "num_base_bdevs_discovered": 0, 00:10:34.853 "num_base_bdevs_operational": 4, 00:10:34.853 "base_bdevs_list": [ 00:10:34.853 { 00:10:34.853 "name": "BaseBdev1", 00:10:34.853 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:34.853 "is_configured": false, 00:10:34.853 "data_offset": 0, 00:10:34.853 "data_size": 0 00:10:34.853 }, 00:10:34.853 { 00:10:34.853 "name": "BaseBdev2", 00:10:34.853 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:34.853 "is_configured": false, 00:10:34.853 "data_offset": 0, 00:10:34.853 "data_size": 0 00:10:34.853 }, 00:10:34.853 { 00:10:34.853 "name": "BaseBdev3", 00:10:34.853 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:34.853 "is_configured": false, 00:10:34.853 "data_offset": 0, 00:10:34.853 "data_size": 0 00:10:34.853 }, 00:10:34.853 { 00:10:34.853 "name": "BaseBdev4", 00:10:34.853 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:34.853 "is_configured": false, 00:10:34.853 "data_offset": 0, 00:10:34.853 "data_size": 0 00:10:34.853 } 00:10:34.853 ] 00:10:34.853 }' 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:34.853 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.424 [2024-12-05 19:00:52.743663] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:35.424 [2024-12-05 19:00:52.743768] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.424 [2024-12-05 19:00:52.755681] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:35.424 [2024-12-05 19:00:52.755775] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:35.424 [2024-12-05 19:00:52.755809] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:35.424 [2024-12-05 19:00:52.755832] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:35.424 [2024-12-05 19:00:52.755850] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:35.424 [2024-12-05 19:00:52.755871] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:35.424 [2024-12-05 19:00:52.755888] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:35.424 [2024-12-05 19:00:52.755908] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.424 [2024-12-05 19:00:52.776564] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:35.424 BaseBdev1 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.424 [ 00:10:35.424 { 00:10:35.424 "name": "BaseBdev1", 00:10:35.424 "aliases": [ 00:10:35.424 "53f0f83c-d93d-47c4-8baf-18b8a7d14950" 00:10:35.424 ], 00:10:35.424 "product_name": "Malloc disk", 00:10:35.424 "block_size": 512, 00:10:35.424 "num_blocks": 65536, 00:10:35.424 "uuid": "53f0f83c-d93d-47c4-8baf-18b8a7d14950", 00:10:35.424 "assigned_rate_limits": { 00:10:35.424 "rw_ios_per_sec": 0, 00:10:35.424 "rw_mbytes_per_sec": 0, 00:10:35.424 "r_mbytes_per_sec": 0, 00:10:35.424 "w_mbytes_per_sec": 0 00:10:35.424 }, 00:10:35.424 "claimed": true, 00:10:35.424 "claim_type": "exclusive_write", 00:10:35.424 "zoned": false, 00:10:35.424 "supported_io_types": { 00:10:35.424 "read": true, 00:10:35.424 "write": true, 00:10:35.424 "unmap": true, 00:10:35.424 "flush": true, 00:10:35.424 "reset": true, 00:10:35.424 "nvme_admin": false, 00:10:35.424 "nvme_io": false, 00:10:35.424 "nvme_io_md": false, 00:10:35.424 "write_zeroes": true, 00:10:35.424 "zcopy": true, 00:10:35.424 "get_zone_info": false, 00:10:35.424 "zone_management": false, 00:10:35.424 "zone_append": false, 00:10:35.424 "compare": false, 00:10:35.424 "compare_and_write": false, 00:10:35.424 "abort": true, 00:10:35.424 "seek_hole": false, 00:10:35.424 "seek_data": false, 00:10:35.424 "copy": true, 00:10:35.424 "nvme_iov_md": false 00:10:35.424 }, 00:10:35.424 "memory_domains": [ 00:10:35.424 { 00:10:35.424 "dma_device_id": "system", 00:10:35.424 "dma_device_type": 1 00:10:35.424 }, 00:10:35.424 { 00:10:35.424 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:35.424 "dma_device_type": 2 00:10:35.424 } 00:10:35.424 ], 00:10:35.424 "driver_specific": {} 00:10:35.424 } 00:10:35.424 ] 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.424 "name": "Existed_Raid", 00:10:35.424 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.424 "strip_size_kb": 0, 00:10:35.424 "state": "configuring", 00:10:35.424 "raid_level": "raid1", 00:10:35.424 "superblock": false, 00:10:35.424 "num_base_bdevs": 4, 00:10:35.424 "num_base_bdevs_discovered": 1, 00:10:35.424 "num_base_bdevs_operational": 4, 00:10:35.424 "base_bdevs_list": [ 00:10:35.424 { 00:10:35.424 "name": "BaseBdev1", 00:10:35.424 "uuid": "53f0f83c-d93d-47c4-8baf-18b8a7d14950", 00:10:35.424 "is_configured": true, 00:10:35.424 "data_offset": 0, 00:10:35.424 "data_size": 65536 00:10:35.424 }, 00:10:35.424 { 00:10:35.424 "name": "BaseBdev2", 00:10:35.424 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.424 "is_configured": false, 00:10:35.424 "data_offset": 0, 00:10:35.424 "data_size": 0 00:10:35.424 }, 00:10:35.424 { 00:10:35.424 "name": "BaseBdev3", 00:10:35.424 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.424 "is_configured": false, 00:10:35.424 "data_offset": 0, 00:10:35.424 "data_size": 0 00:10:35.424 }, 00:10:35.424 { 00:10:35.424 "name": "BaseBdev4", 00:10:35.424 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.424 "is_configured": false, 00:10:35.424 "data_offset": 0, 00:10:35.424 "data_size": 0 00:10:35.424 } 00:10:35.424 ] 00:10:35.424 }' 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.424 19:00:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.684 [2024-12-05 19:00:53.215862] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:35.684 [2024-12-05 19:00:53.215945] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.684 [2024-12-05 19:00:53.227876] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:35.684 [2024-12-05 19:00:53.229820] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:35.684 [2024-12-05 19:00:53.229912] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:35.684 [2024-12-05 19:00:53.229940] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:35.684 [2024-12-05 19:00:53.229962] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:35.684 [2024-12-05 19:00:53.229979] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:35.684 [2024-12-05 19:00:53.229999] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:35.684 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.944 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:35.944 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.944 "name": "Existed_Raid", 00:10:35.944 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.944 "strip_size_kb": 0, 00:10:35.944 "state": "configuring", 00:10:35.944 "raid_level": "raid1", 00:10:35.944 "superblock": false, 00:10:35.944 "num_base_bdevs": 4, 00:10:35.944 "num_base_bdevs_discovered": 1, 00:10:35.944 "num_base_bdevs_operational": 4, 00:10:35.944 "base_bdevs_list": [ 00:10:35.944 { 00:10:35.944 "name": "BaseBdev1", 00:10:35.944 "uuid": "53f0f83c-d93d-47c4-8baf-18b8a7d14950", 00:10:35.944 "is_configured": true, 00:10:35.944 "data_offset": 0, 00:10:35.944 "data_size": 65536 00:10:35.944 }, 00:10:35.944 { 00:10:35.944 "name": "BaseBdev2", 00:10:35.944 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.944 "is_configured": false, 00:10:35.944 "data_offset": 0, 00:10:35.944 "data_size": 0 00:10:35.944 }, 00:10:35.944 { 00:10:35.944 "name": "BaseBdev3", 00:10:35.944 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.944 "is_configured": false, 00:10:35.944 "data_offset": 0, 00:10:35.944 "data_size": 0 00:10:35.944 }, 00:10:35.944 { 00:10:35.944 "name": "BaseBdev4", 00:10:35.944 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.944 "is_configured": false, 00:10:35.944 "data_offset": 0, 00:10:35.944 "data_size": 0 00:10:35.944 } 00:10:35.944 ] 00:10:35.944 }' 00:10:35.944 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.944 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.204 [2024-12-05 19:00:53.638218] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:36.204 BaseBdev2 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.204 [ 00:10:36.204 { 00:10:36.204 "name": "BaseBdev2", 00:10:36.204 "aliases": [ 00:10:36.204 "66946609-f465-48b2-9644-cd7c33815490" 00:10:36.204 ], 00:10:36.204 "product_name": "Malloc disk", 00:10:36.204 "block_size": 512, 00:10:36.204 "num_blocks": 65536, 00:10:36.204 "uuid": "66946609-f465-48b2-9644-cd7c33815490", 00:10:36.204 "assigned_rate_limits": { 00:10:36.204 "rw_ios_per_sec": 0, 00:10:36.204 "rw_mbytes_per_sec": 0, 00:10:36.204 "r_mbytes_per_sec": 0, 00:10:36.204 "w_mbytes_per_sec": 0 00:10:36.204 }, 00:10:36.204 "claimed": true, 00:10:36.204 "claim_type": "exclusive_write", 00:10:36.204 "zoned": false, 00:10:36.204 "supported_io_types": { 00:10:36.204 "read": true, 00:10:36.204 "write": true, 00:10:36.204 "unmap": true, 00:10:36.204 "flush": true, 00:10:36.204 "reset": true, 00:10:36.204 "nvme_admin": false, 00:10:36.204 "nvme_io": false, 00:10:36.204 "nvme_io_md": false, 00:10:36.204 "write_zeroes": true, 00:10:36.204 "zcopy": true, 00:10:36.204 "get_zone_info": false, 00:10:36.204 "zone_management": false, 00:10:36.204 "zone_append": false, 00:10:36.204 "compare": false, 00:10:36.204 "compare_and_write": false, 00:10:36.204 "abort": true, 00:10:36.204 "seek_hole": false, 00:10:36.204 "seek_data": false, 00:10:36.204 "copy": true, 00:10:36.204 "nvme_iov_md": false 00:10:36.204 }, 00:10:36.204 "memory_domains": [ 00:10:36.204 { 00:10:36.204 "dma_device_id": "system", 00:10:36.204 "dma_device_type": 1 00:10:36.204 }, 00:10:36.204 { 00:10:36.204 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:36.204 "dma_device_type": 2 00:10:36.204 } 00:10:36.204 ], 00:10:36.204 "driver_specific": {} 00:10:36.204 } 00:10:36.204 ] 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:36.204 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:36.205 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:36.205 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.205 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.205 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.205 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.205 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:36.205 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.205 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.205 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.205 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.205 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.205 "name": "Existed_Raid", 00:10:36.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:36.205 "strip_size_kb": 0, 00:10:36.205 "state": "configuring", 00:10:36.205 "raid_level": "raid1", 00:10:36.205 "superblock": false, 00:10:36.205 "num_base_bdevs": 4, 00:10:36.205 "num_base_bdevs_discovered": 2, 00:10:36.205 "num_base_bdevs_operational": 4, 00:10:36.205 "base_bdevs_list": [ 00:10:36.205 { 00:10:36.205 "name": "BaseBdev1", 00:10:36.205 "uuid": "53f0f83c-d93d-47c4-8baf-18b8a7d14950", 00:10:36.205 "is_configured": true, 00:10:36.205 "data_offset": 0, 00:10:36.205 "data_size": 65536 00:10:36.205 }, 00:10:36.205 { 00:10:36.205 "name": "BaseBdev2", 00:10:36.205 "uuid": "66946609-f465-48b2-9644-cd7c33815490", 00:10:36.205 "is_configured": true, 00:10:36.205 "data_offset": 0, 00:10:36.205 "data_size": 65536 00:10:36.205 }, 00:10:36.205 { 00:10:36.205 "name": "BaseBdev3", 00:10:36.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:36.205 "is_configured": false, 00:10:36.205 "data_offset": 0, 00:10:36.205 "data_size": 0 00:10:36.205 }, 00:10:36.205 { 00:10:36.205 "name": "BaseBdev4", 00:10:36.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:36.205 "is_configured": false, 00:10:36.205 "data_offset": 0, 00:10:36.205 "data_size": 0 00:10:36.205 } 00:10:36.205 ] 00:10:36.205 }' 00:10:36.205 19:00:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.205 19:00:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.775 [2024-12-05 19:00:54.081844] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:36.775 BaseBdev3 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.775 [ 00:10:36.775 { 00:10:36.775 "name": "BaseBdev3", 00:10:36.775 "aliases": [ 00:10:36.775 "7ac03ea4-f365-47c0-8243-6f78b2c59bc7" 00:10:36.775 ], 00:10:36.775 "product_name": "Malloc disk", 00:10:36.775 "block_size": 512, 00:10:36.775 "num_blocks": 65536, 00:10:36.775 "uuid": "7ac03ea4-f365-47c0-8243-6f78b2c59bc7", 00:10:36.775 "assigned_rate_limits": { 00:10:36.775 "rw_ios_per_sec": 0, 00:10:36.775 "rw_mbytes_per_sec": 0, 00:10:36.775 "r_mbytes_per_sec": 0, 00:10:36.775 "w_mbytes_per_sec": 0 00:10:36.775 }, 00:10:36.775 "claimed": true, 00:10:36.775 "claim_type": "exclusive_write", 00:10:36.775 "zoned": false, 00:10:36.775 "supported_io_types": { 00:10:36.775 "read": true, 00:10:36.775 "write": true, 00:10:36.775 "unmap": true, 00:10:36.775 "flush": true, 00:10:36.775 "reset": true, 00:10:36.775 "nvme_admin": false, 00:10:36.775 "nvme_io": false, 00:10:36.775 "nvme_io_md": false, 00:10:36.775 "write_zeroes": true, 00:10:36.775 "zcopy": true, 00:10:36.775 "get_zone_info": false, 00:10:36.775 "zone_management": false, 00:10:36.775 "zone_append": false, 00:10:36.775 "compare": false, 00:10:36.775 "compare_and_write": false, 00:10:36.775 "abort": true, 00:10:36.775 "seek_hole": false, 00:10:36.775 "seek_data": false, 00:10:36.775 "copy": true, 00:10:36.775 "nvme_iov_md": false 00:10:36.775 }, 00:10:36.775 "memory_domains": [ 00:10:36.775 { 00:10:36.775 "dma_device_id": "system", 00:10:36.775 "dma_device_type": 1 00:10:36.775 }, 00:10:36.775 { 00:10:36.775 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:36.775 "dma_device_type": 2 00:10:36.775 } 00:10:36.775 ], 00:10:36.775 "driver_specific": {} 00:10:36.775 } 00:10:36.775 ] 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.775 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.776 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:36.776 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:36.776 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.776 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:36.776 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.776 "name": "Existed_Raid", 00:10:36.776 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:36.776 "strip_size_kb": 0, 00:10:36.776 "state": "configuring", 00:10:36.776 "raid_level": "raid1", 00:10:36.776 "superblock": false, 00:10:36.776 "num_base_bdevs": 4, 00:10:36.776 "num_base_bdevs_discovered": 3, 00:10:36.776 "num_base_bdevs_operational": 4, 00:10:36.776 "base_bdevs_list": [ 00:10:36.776 { 00:10:36.776 "name": "BaseBdev1", 00:10:36.776 "uuid": "53f0f83c-d93d-47c4-8baf-18b8a7d14950", 00:10:36.776 "is_configured": true, 00:10:36.776 "data_offset": 0, 00:10:36.776 "data_size": 65536 00:10:36.776 }, 00:10:36.776 { 00:10:36.776 "name": "BaseBdev2", 00:10:36.776 "uuid": "66946609-f465-48b2-9644-cd7c33815490", 00:10:36.776 "is_configured": true, 00:10:36.776 "data_offset": 0, 00:10:36.776 "data_size": 65536 00:10:36.776 }, 00:10:36.776 { 00:10:36.776 "name": "BaseBdev3", 00:10:36.776 "uuid": "7ac03ea4-f365-47c0-8243-6f78b2c59bc7", 00:10:36.776 "is_configured": true, 00:10:36.776 "data_offset": 0, 00:10:36.776 "data_size": 65536 00:10:36.776 }, 00:10:36.776 { 00:10:36.776 "name": "BaseBdev4", 00:10:36.776 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:36.776 "is_configured": false, 00:10:36.776 "data_offset": 0, 00:10:36.776 "data_size": 0 00:10:36.776 } 00:10:36.776 ] 00:10:36.776 }' 00:10:36.776 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.776 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.036 [2024-12-05 19:00:54.544097] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:37.036 [2024-12-05 19:00:54.544224] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:37.036 [2024-12-05 19:00:54.544250] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:37.036 [2024-12-05 19:00:54.544603] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:37.036 [2024-12-05 19:00:54.544815] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:37.036 BaseBdev4 00:10:37.036 [2024-12-05 19:00:54.544869] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:37.036 [2024-12-05 19:00:54.545075] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.036 [ 00:10:37.036 { 00:10:37.036 "name": "BaseBdev4", 00:10:37.036 "aliases": [ 00:10:37.036 "d897f2b5-5d57-461e-b254-482561cfbfc4" 00:10:37.036 ], 00:10:37.036 "product_name": "Malloc disk", 00:10:37.036 "block_size": 512, 00:10:37.036 "num_blocks": 65536, 00:10:37.036 "uuid": "d897f2b5-5d57-461e-b254-482561cfbfc4", 00:10:37.036 "assigned_rate_limits": { 00:10:37.036 "rw_ios_per_sec": 0, 00:10:37.036 "rw_mbytes_per_sec": 0, 00:10:37.036 "r_mbytes_per_sec": 0, 00:10:37.036 "w_mbytes_per_sec": 0 00:10:37.036 }, 00:10:37.036 "claimed": true, 00:10:37.036 "claim_type": "exclusive_write", 00:10:37.036 "zoned": false, 00:10:37.036 "supported_io_types": { 00:10:37.036 "read": true, 00:10:37.036 "write": true, 00:10:37.036 "unmap": true, 00:10:37.036 "flush": true, 00:10:37.036 "reset": true, 00:10:37.036 "nvme_admin": false, 00:10:37.036 "nvme_io": false, 00:10:37.036 "nvme_io_md": false, 00:10:37.036 "write_zeroes": true, 00:10:37.036 "zcopy": true, 00:10:37.036 "get_zone_info": false, 00:10:37.036 "zone_management": false, 00:10:37.036 "zone_append": false, 00:10:37.036 "compare": false, 00:10:37.036 "compare_and_write": false, 00:10:37.036 "abort": true, 00:10:37.036 "seek_hole": false, 00:10:37.036 "seek_data": false, 00:10:37.036 "copy": true, 00:10:37.036 "nvme_iov_md": false 00:10:37.036 }, 00:10:37.036 "memory_domains": [ 00:10:37.036 { 00:10:37.036 "dma_device_id": "system", 00:10:37.036 "dma_device_type": 1 00:10:37.036 }, 00:10:37.036 { 00:10:37.036 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.036 "dma_device_type": 2 00:10:37.036 } 00:10:37.036 ], 00:10:37.036 "driver_specific": {} 00:10:37.036 } 00:10:37.036 ] 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.036 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.296 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.296 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:37.296 "name": "Existed_Raid", 00:10:37.296 "uuid": "1407b12f-6b7f-4ca2-a621-5341983a1040", 00:10:37.296 "strip_size_kb": 0, 00:10:37.296 "state": "online", 00:10:37.296 "raid_level": "raid1", 00:10:37.296 "superblock": false, 00:10:37.296 "num_base_bdevs": 4, 00:10:37.296 "num_base_bdevs_discovered": 4, 00:10:37.296 "num_base_bdevs_operational": 4, 00:10:37.296 "base_bdevs_list": [ 00:10:37.296 { 00:10:37.296 "name": "BaseBdev1", 00:10:37.296 "uuid": "53f0f83c-d93d-47c4-8baf-18b8a7d14950", 00:10:37.296 "is_configured": true, 00:10:37.296 "data_offset": 0, 00:10:37.296 "data_size": 65536 00:10:37.296 }, 00:10:37.296 { 00:10:37.296 "name": "BaseBdev2", 00:10:37.296 "uuid": "66946609-f465-48b2-9644-cd7c33815490", 00:10:37.296 "is_configured": true, 00:10:37.296 "data_offset": 0, 00:10:37.296 "data_size": 65536 00:10:37.296 }, 00:10:37.296 { 00:10:37.296 "name": "BaseBdev3", 00:10:37.296 "uuid": "7ac03ea4-f365-47c0-8243-6f78b2c59bc7", 00:10:37.296 "is_configured": true, 00:10:37.296 "data_offset": 0, 00:10:37.296 "data_size": 65536 00:10:37.296 }, 00:10:37.296 { 00:10:37.296 "name": "BaseBdev4", 00:10:37.296 "uuid": "d897f2b5-5d57-461e-b254-482561cfbfc4", 00:10:37.296 "is_configured": true, 00:10:37.296 "data_offset": 0, 00:10:37.296 "data_size": 65536 00:10:37.296 } 00:10:37.296 ] 00:10:37.296 }' 00:10:37.297 19:00:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:37.297 19:00:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.557 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:37.557 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:37.557 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:37.557 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:37.557 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:37.557 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:37.557 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:37.557 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.557 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.557 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:37.557 [2024-12-05 19:00:55.015620] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:37.557 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.557 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:37.557 "name": "Existed_Raid", 00:10:37.557 "aliases": [ 00:10:37.557 "1407b12f-6b7f-4ca2-a621-5341983a1040" 00:10:37.557 ], 00:10:37.557 "product_name": "Raid Volume", 00:10:37.557 "block_size": 512, 00:10:37.557 "num_blocks": 65536, 00:10:37.557 "uuid": "1407b12f-6b7f-4ca2-a621-5341983a1040", 00:10:37.557 "assigned_rate_limits": { 00:10:37.557 "rw_ios_per_sec": 0, 00:10:37.557 "rw_mbytes_per_sec": 0, 00:10:37.557 "r_mbytes_per_sec": 0, 00:10:37.557 "w_mbytes_per_sec": 0 00:10:37.557 }, 00:10:37.557 "claimed": false, 00:10:37.557 "zoned": false, 00:10:37.557 "supported_io_types": { 00:10:37.557 "read": true, 00:10:37.557 "write": true, 00:10:37.557 "unmap": false, 00:10:37.557 "flush": false, 00:10:37.557 "reset": true, 00:10:37.557 "nvme_admin": false, 00:10:37.557 "nvme_io": false, 00:10:37.557 "nvme_io_md": false, 00:10:37.557 "write_zeroes": true, 00:10:37.557 "zcopy": false, 00:10:37.557 "get_zone_info": false, 00:10:37.557 "zone_management": false, 00:10:37.557 "zone_append": false, 00:10:37.557 "compare": false, 00:10:37.557 "compare_and_write": false, 00:10:37.557 "abort": false, 00:10:37.557 "seek_hole": false, 00:10:37.557 "seek_data": false, 00:10:37.557 "copy": false, 00:10:37.557 "nvme_iov_md": false 00:10:37.557 }, 00:10:37.557 "memory_domains": [ 00:10:37.557 { 00:10:37.557 "dma_device_id": "system", 00:10:37.557 "dma_device_type": 1 00:10:37.557 }, 00:10:37.557 { 00:10:37.557 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.557 "dma_device_type": 2 00:10:37.557 }, 00:10:37.557 { 00:10:37.557 "dma_device_id": "system", 00:10:37.557 "dma_device_type": 1 00:10:37.557 }, 00:10:37.557 { 00:10:37.557 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.557 "dma_device_type": 2 00:10:37.557 }, 00:10:37.557 { 00:10:37.557 "dma_device_id": "system", 00:10:37.557 "dma_device_type": 1 00:10:37.557 }, 00:10:37.557 { 00:10:37.557 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.557 "dma_device_type": 2 00:10:37.557 }, 00:10:37.557 { 00:10:37.557 "dma_device_id": "system", 00:10:37.557 "dma_device_type": 1 00:10:37.557 }, 00:10:37.557 { 00:10:37.557 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.557 "dma_device_type": 2 00:10:37.557 } 00:10:37.557 ], 00:10:37.557 "driver_specific": { 00:10:37.557 "raid": { 00:10:37.557 "uuid": "1407b12f-6b7f-4ca2-a621-5341983a1040", 00:10:37.557 "strip_size_kb": 0, 00:10:37.557 "state": "online", 00:10:37.557 "raid_level": "raid1", 00:10:37.557 "superblock": false, 00:10:37.557 "num_base_bdevs": 4, 00:10:37.557 "num_base_bdevs_discovered": 4, 00:10:37.557 "num_base_bdevs_operational": 4, 00:10:37.557 "base_bdevs_list": [ 00:10:37.557 { 00:10:37.557 "name": "BaseBdev1", 00:10:37.557 "uuid": "53f0f83c-d93d-47c4-8baf-18b8a7d14950", 00:10:37.557 "is_configured": true, 00:10:37.557 "data_offset": 0, 00:10:37.557 "data_size": 65536 00:10:37.557 }, 00:10:37.557 { 00:10:37.557 "name": "BaseBdev2", 00:10:37.557 "uuid": "66946609-f465-48b2-9644-cd7c33815490", 00:10:37.557 "is_configured": true, 00:10:37.557 "data_offset": 0, 00:10:37.557 "data_size": 65536 00:10:37.557 }, 00:10:37.557 { 00:10:37.557 "name": "BaseBdev3", 00:10:37.557 "uuid": "7ac03ea4-f365-47c0-8243-6f78b2c59bc7", 00:10:37.557 "is_configured": true, 00:10:37.557 "data_offset": 0, 00:10:37.557 "data_size": 65536 00:10:37.557 }, 00:10:37.557 { 00:10:37.557 "name": "BaseBdev4", 00:10:37.557 "uuid": "d897f2b5-5d57-461e-b254-482561cfbfc4", 00:10:37.557 "is_configured": true, 00:10:37.557 "data_offset": 0, 00:10:37.557 "data_size": 65536 00:10:37.557 } 00:10:37.557 ] 00:10:37.557 } 00:10:37.557 } 00:10:37.557 }' 00:10:37.557 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:37.557 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:37.557 BaseBdev2 00:10:37.557 BaseBdev3 00:10:37.557 BaseBdev4' 00:10:37.557 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.817 [2024-12-05 19:00:55.338799] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:37.817 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.077 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.077 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:38.077 "name": "Existed_Raid", 00:10:38.077 "uuid": "1407b12f-6b7f-4ca2-a621-5341983a1040", 00:10:38.077 "strip_size_kb": 0, 00:10:38.077 "state": "online", 00:10:38.077 "raid_level": "raid1", 00:10:38.077 "superblock": false, 00:10:38.077 "num_base_bdevs": 4, 00:10:38.077 "num_base_bdevs_discovered": 3, 00:10:38.077 "num_base_bdevs_operational": 3, 00:10:38.077 "base_bdevs_list": [ 00:10:38.077 { 00:10:38.077 "name": null, 00:10:38.077 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:38.077 "is_configured": false, 00:10:38.077 "data_offset": 0, 00:10:38.077 "data_size": 65536 00:10:38.077 }, 00:10:38.077 { 00:10:38.077 "name": "BaseBdev2", 00:10:38.077 "uuid": "66946609-f465-48b2-9644-cd7c33815490", 00:10:38.077 "is_configured": true, 00:10:38.077 "data_offset": 0, 00:10:38.077 "data_size": 65536 00:10:38.077 }, 00:10:38.077 { 00:10:38.077 "name": "BaseBdev3", 00:10:38.077 "uuid": "7ac03ea4-f365-47c0-8243-6f78b2c59bc7", 00:10:38.077 "is_configured": true, 00:10:38.077 "data_offset": 0, 00:10:38.077 "data_size": 65536 00:10:38.077 }, 00:10:38.077 { 00:10:38.077 "name": "BaseBdev4", 00:10:38.077 "uuid": "d897f2b5-5d57-461e-b254-482561cfbfc4", 00:10:38.077 "is_configured": true, 00:10:38.077 "data_offset": 0, 00:10:38.077 "data_size": 65536 00:10:38.077 } 00:10:38.077 ] 00:10:38.077 }' 00:10:38.077 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:38.077 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.338 [2024-12-05 19:00:55.833271] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.338 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.598 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:38.598 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:38.598 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:38.598 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.598 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.598 [2024-12-05 19:00:55.904251] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.599 [2024-12-05 19:00:55.971175] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:38.599 [2024-12-05 19:00:55.971301] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:38.599 [2024-12-05 19:00:55.982500] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:38.599 [2024-12-05 19:00:55.982610] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:38.599 [2024-12-05 19:00:55.982651] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.599 19:00:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.599 BaseBdev2 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.599 [ 00:10:38.599 { 00:10:38.599 "name": "BaseBdev2", 00:10:38.599 "aliases": [ 00:10:38.599 "e6876914-afda-4401-93b5-5a1e2448fe8f" 00:10:38.599 ], 00:10:38.599 "product_name": "Malloc disk", 00:10:38.599 "block_size": 512, 00:10:38.599 "num_blocks": 65536, 00:10:38.599 "uuid": "e6876914-afda-4401-93b5-5a1e2448fe8f", 00:10:38.599 "assigned_rate_limits": { 00:10:38.599 "rw_ios_per_sec": 0, 00:10:38.599 "rw_mbytes_per_sec": 0, 00:10:38.599 "r_mbytes_per_sec": 0, 00:10:38.599 "w_mbytes_per_sec": 0 00:10:38.599 }, 00:10:38.599 "claimed": false, 00:10:38.599 "zoned": false, 00:10:38.599 "supported_io_types": { 00:10:38.599 "read": true, 00:10:38.599 "write": true, 00:10:38.599 "unmap": true, 00:10:38.599 "flush": true, 00:10:38.599 "reset": true, 00:10:38.599 "nvme_admin": false, 00:10:38.599 "nvme_io": false, 00:10:38.599 "nvme_io_md": false, 00:10:38.599 "write_zeroes": true, 00:10:38.599 "zcopy": true, 00:10:38.599 "get_zone_info": false, 00:10:38.599 "zone_management": false, 00:10:38.599 "zone_append": false, 00:10:38.599 "compare": false, 00:10:38.599 "compare_and_write": false, 00:10:38.599 "abort": true, 00:10:38.599 "seek_hole": false, 00:10:38.599 "seek_data": false, 00:10:38.599 "copy": true, 00:10:38.599 "nvme_iov_md": false 00:10:38.599 }, 00:10:38.599 "memory_domains": [ 00:10:38.599 { 00:10:38.599 "dma_device_id": "system", 00:10:38.599 "dma_device_type": 1 00:10:38.599 }, 00:10:38.599 { 00:10:38.599 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:38.599 "dma_device_type": 2 00:10:38.599 } 00:10:38.599 ], 00:10:38.599 "driver_specific": {} 00:10:38.599 } 00:10:38.599 ] 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.599 BaseBdev3 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:38.599 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.600 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.600 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.600 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:38.600 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.600 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.600 [ 00:10:38.600 { 00:10:38.600 "name": "BaseBdev3", 00:10:38.600 "aliases": [ 00:10:38.600 "2cf8b627-d211-4fdf-8173-4c8758dba086" 00:10:38.600 ], 00:10:38.600 "product_name": "Malloc disk", 00:10:38.600 "block_size": 512, 00:10:38.600 "num_blocks": 65536, 00:10:38.600 "uuid": "2cf8b627-d211-4fdf-8173-4c8758dba086", 00:10:38.600 "assigned_rate_limits": { 00:10:38.600 "rw_ios_per_sec": 0, 00:10:38.600 "rw_mbytes_per_sec": 0, 00:10:38.600 "r_mbytes_per_sec": 0, 00:10:38.600 "w_mbytes_per_sec": 0 00:10:38.600 }, 00:10:38.600 "claimed": false, 00:10:38.600 "zoned": false, 00:10:38.600 "supported_io_types": { 00:10:38.600 "read": true, 00:10:38.600 "write": true, 00:10:38.600 "unmap": true, 00:10:38.600 "flush": true, 00:10:38.600 "reset": true, 00:10:38.600 "nvme_admin": false, 00:10:38.600 "nvme_io": false, 00:10:38.600 "nvme_io_md": false, 00:10:38.600 "write_zeroes": true, 00:10:38.600 "zcopy": true, 00:10:38.600 "get_zone_info": false, 00:10:38.600 "zone_management": false, 00:10:38.600 "zone_append": false, 00:10:38.600 "compare": false, 00:10:38.600 "compare_and_write": false, 00:10:38.600 "abort": true, 00:10:38.600 "seek_hole": false, 00:10:38.600 "seek_data": false, 00:10:38.600 "copy": true, 00:10:38.600 "nvme_iov_md": false 00:10:38.600 }, 00:10:38.600 "memory_domains": [ 00:10:38.600 { 00:10:38.600 "dma_device_id": "system", 00:10:38.600 "dma_device_type": 1 00:10:38.600 }, 00:10:38.600 { 00:10:38.600 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:38.600 "dma_device_type": 2 00:10:38.600 } 00:10:38.600 ], 00:10:38.600 "driver_specific": {} 00:10:38.600 } 00:10:38.600 ] 00:10:38.600 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.600 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:38.600 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:38.600 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:38.600 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:38.600 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.600 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.600 BaseBdev4 00:10:38.600 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.600 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:38.600 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.861 [ 00:10:38.861 { 00:10:38.861 "name": "BaseBdev4", 00:10:38.861 "aliases": [ 00:10:38.861 "57ca4575-dbdd-4ac7-aba0-11c851287ee6" 00:10:38.861 ], 00:10:38.861 "product_name": "Malloc disk", 00:10:38.861 "block_size": 512, 00:10:38.861 "num_blocks": 65536, 00:10:38.861 "uuid": "57ca4575-dbdd-4ac7-aba0-11c851287ee6", 00:10:38.861 "assigned_rate_limits": { 00:10:38.861 "rw_ios_per_sec": 0, 00:10:38.861 "rw_mbytes_per_sec": 0, 00:10:38.861 "r_mbytes_per_sec": 0, 00:10:38.861 "w_mbytes_per_sec": 0 00:10:38.861 }, 00:10:38.861 "claimed": false, 00:10:38.861 "zoned": false, 00:10:38.861 "supported_io_types": { 00:10:38.861 "read": true, 00:10:38.861 "write": true, 00:10:38.861 "unmap": true, 00:10:38.861 "flush": true, 00:10:38.861 "reset": true, 00:10:38.861 "nvme_admin": false, 00:10:38.861 "nvme_io": false, 00:10:38.861 "nvme_io_md": false, 00:10:38.861 "write_zeroes": true, 00:10:38.861 "zcopy": true, 00:10:38.861 "get_zone_info": false, 00:10:38.861 "zone_management": false, 00:10:38.861 "zone_append": false, 00:10:38.861 "compare": false, 00:10:38.861 "compare_and_write": false, 00:10:38.861 "abort": true, 00:10:38.861 "seek_hole": false, 00:10:38.861 "seek_data": false, 00:10:38.861 "copy": true, 00:10:38.861 "nvme_iov_md": false 00:10:38.861 }, 00:10:38.861 "memory_domains": [ 00:10:38.861 { 00:10:38.861 "dma_device_id": "system", 00:10:38.861 "dma_device_type": 1 00:10:38.861 }, 00:10:38.861 { 00:10:38.861 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:38.861 "dma_device_type": 2 00:10:38.861 } 00:10:38.861 ], 00:10:38.861 "driver_specific": {} 00:10:38.861 } 00:10:38.861 ] 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.861 [2024-12-05 19:00:56.198455] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:38.861 [2024-12-05 19:00:56.198562] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:38.861 [2024-12-05 19:00:56.198604] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:38.861 [2024-12-05 19:00:56.200428] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:38.861 [2024-12-05 19:00:56.200539] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:38.861 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:38.861 "name": "Existed_Raid", 00:10:38.861 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:38.861 "strip_size_kb": 0, 00:10:38.861 "state": "configuring", 00:10:38.861 "raid_level": "raid1", 00:10:38.861 "superblock": false, 00:10:38.861 "num_base_bdevs": 4, 00:10:38.861 "num_base_bdevs_discovered": 3, 00:10:38.861 "num_base_bdevs_operational": 4, 00:10:38.861 "base_bdevs_list": [ 00:10:38.861 { 00:10:38.861 "name": "BaseBdev1", 00:10:38.861 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:38.861 "is_configured": false, 00:10:38.861 "data_offset": 0, 00:10:38.861 "data_size": 0 00:10:38.861 }, 00:10:38.861 { 00:10:38.861 "name": "BaseBdev2", 00:10:38.861 "uuid": "e6876914-afda-4401-93b5-5a1e2448fe8f", 00:10:38.861 "is_configured": true, 00:10:38.861 "data_offset": 0, 00:10:38.861 "data_size": 65536 00:10:38.862 }, 00:10:38.862 { 00:10:38.862 "name": "BaseBdev3", 00:10:38.862 "uuid": "2cf8b627-d211-4fdf-8173-4c8758dba086", 00:10:38.862 "is_configured": true, 00:10:38.862 "data_offset": 0, 00:10:38.862 "data_size": 65536 00:10:38.862 }, 00:10:38.862 { 00:10:38.862 "name": "BaseBdev4", 00:10:38.862 "uuid": "57ca4575-dbdd-4ac7-aba0-11c851287ee6", 00:10:38.862 "is_configured": true, 00:10:38.862 "data_offset": 0, 00:10:38.862 "data_size": 65536 00:10:38.862 } 00:10:38.862 ] 00:10:38.862 }' 00:10:38.862 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:38.862 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.122 [2024-12-05 19:00:56.657680] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.122 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.382 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.383 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:39.383 "name": "Existed_Raid", 00:10:39.383 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.383 "strip_size_kb": 0, 00:10:39.383 "state": "configuring", 00:10:39.383 "raid_level": "raid1", 00:10:39.383 "superblock": false, 00:10:39.383 "num_base_bdevs": 4, 00:10:39.383 "num_base_bdevs_discovered": 2, 00:10:39.383 "num_base_bdevs_operational": 4, 00:10:39.383 "base_bdevs_list": [ 00:10:39.383 { 00:10:39.383 "name": "BaseBdev1", 00:10:39.383 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.383 "is_configured": false, 00:10:39.383 "data_offset": 0, 00:10:39.383 "data_size": 0 00:10:39.383 }, 00:10:39.383 { 00:10:39.383 "name": null, 00:10:39.383 "uuid": "e6876914-afda-4401-93b5-5a1e2448fe8f", 00:10:39.383 "is_configured": false, 00:10:39.383 "data_offset": 0, 00:10:39.383 "data_size": 65536 00:10:39.383 }, 00:10:39.383 { 00:10:39.383 "name": "BaseBdev3", 00:10:39.383 "uuid": "2cf8b627-d211-4fdf-8173-4c8758dba086", 00:10:39.383 "is_configured": true, 00:10:39.383 "data_offset": 0, 00:10:39.383 "data_size": 65536 00:10:39.383 }, 00:10:39.383 { 00:10:39.383 "name": "BaseBdev4", 00:10:39.383 "uuid": "57ca4575-dbdd-4ac7-aba0-11c851287ee6", 00:10:39.383 "is_configured": true, 00:10:39.383 "data_offset": 0, 00:10:39.383 "data_size": 65536 00:10:39.383 } 00:10:39.383 ] 00:10:39.383 }' 00:10:39.383 19:00:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:39.383 19:00:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.642 [2024-12-05 19:00:57.183876] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:39.642 BaseBdev1 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.642 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:39.643 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.643 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.902 [ 00:10:39.902 { 00:10:39.902 "name": "BaseBdev1", 00:10:39.902 "aliases": [ 00:10:39.902 "bb76236c-502f-4ccf-b213-c4886c9bcffe" 00:10:39.902 ], 00:10:39.902 "product_name": "Malloc disk", 00:10:39.902 "block_size": 512, 00:10:39.902 "num_blocks": 65536, 00:10:39.902 "uuid": "bb76236c-502f-4ccf-b213-c4886c9bcffe", 00:10:39.902 "assigned_rate_limits": { 00:10:39.902 "rw_ios_per_sec": 0, 00:10:39.902 "rw_mbytes_per_sec": 0, 00:10:39.902 "r_mbytes_per_sec": 0, 00:10:39.902 "w_mbytes_per_sec": 0 00:10:39.902 }, 00:10:39.902 "claimed": true, 00:10:39.902 "claim_type": "exclusive_write", 00:10:39.902 "zoned": false, 00:10:39.902 "supported_io_types": { 00:10:39.902 "read": true, 00:10:39.902 "write": true, 00:10:39.902 "unmap": true, 00:10:39.902 "flush": true, 00:10:39.902 "reset": true, 00:10:39.902 "nvme_admin": false, 00:10:39.902 "nvme_io": false, 00:10:39.902 "nvme_io_md": false, 00:10:39.902 "write_zeroes": true, 00:10:39.902 "zcopy": true, 00:10:39.902 "get_zone_info": false, 00:10:39.902 "zone_management": false, 00:10:39.902 "zone_append": false, 00:10:39.902 "compare": false, 00:10:39.902 "compare_and_write": false, 00:10:39.902 "abort": true, 00:10:39.902 "seek_hole": false, 00:10:39.902 "seek_data": false, 00:10:39.902 "copy": true, 00:10:39.902 "nvme_iov_md": false 00:10:39.902 }, 00:10:39.902 "memory_domains": [ 00:10:39.902 { 00:10:39.902 "dma_device_id": "system", 00:10:39.902 "dma_device_type": 1 00:10:39.902 }, 00:10:39.902 { 00:10:39.902 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:39.902 "dma_device_type": 2 00:10:39.902 } 00:10:39.902 ], 00:10:39.902 "driver_specific": {} 00:10:39.902 } 00:10:39.902 ] 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:39.902 "name": "Existed_Raid", 00:10:39.902 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.902 "strip_size_kb": 0, 00:10:39.902 "state": "configuring", 00:10:39.902 "raid_level": "raid1", 00:10:39.902 "superblock": false, 00:10:39.902 "num_base_bdevs": 4, 00:10:39.902 "num_base_bdevs_discovered": 3, 00:10:39.902 "num_base_bdevs_operational": 4, 00:10:39.902 "base_bdevs_list": [ 00:10:39.902 { 00:10:39.902 "name": "BaseBdev1", 00:10:39.902 "uuid": "bb76236c-502f-4ccf-b213-c4886c9bcffe", 00:10:39.902 "is_configured": true, 00:10:39.902 "data_offset": 0, 00:10:39.902 "data_size": 65536 00:10:39.902 }, 00:10:39.902 { 00:10:39.902 "name": null, 00:10:39.902 "uuid": "e6876914-afda-4401-93b5-5a1e2448fe8f", 00:10:39.902 "is_configured": false, 00:10:39.902 "data_offset": 0, 00:10:39.902 "data_size": 65536 00:10:39.902 }, 00:10:39.902 { 00:10:39.902 "name": "BaseBdev3", 00:10:39.902 "uuid": "2cf8b627-d211-4fdf-8173-4c8758dba086", 00:10:39.902 "is_configured": true, 00:10:39.902 "data_offset": 0, 00:10:39.902 "data_size": 65536 00:10:39.902 }, 00:10:39.902 { 00:10:39.902 "name": "BaseBdev4", 00:10:39.902 "uuid": "57ca4575-dbdd-4ac7-aba0-11c851287ee6", 00:10:39.902 "is_configured": true, 00:10:39.902 "data_offset": 0, 00:10:39.902 "data_size": 65536 00:10:39.902 } 00:10:39.902 ] 00:10:39.902 }' 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:39.902 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.162 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.162 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:40.162 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.162 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.162 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.162 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:40.162 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:40.162 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.162 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.162 [2024-12-05 19:00:57.719052] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:40.422 "name": "Existed_Raid", 00:10:40.422 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.422 "strip_size_kb": 0, 00:10:40.422 "state": "configuring", 00:10:40.422 "raid_level": "raid1", 00:10:40.422 "superblock": false, 00:10:40.422 "num_base_bdevs": 4, 00:10:40.422 "num_base_bdevs_discovered": 2, 00:10:40.422 "num_base_bdevs_operational": 4, 00:10:40.422 "base_bdevs_list": [ 00:10:40.422 { 00:10:40.422 "name": "BaseBdev1", 00:10:40.422 "uuid": "bb76236c-502f-4ccf-b213-c4886c9bcffe", 00:10:40.422 "is_configured": true, 00:10:40.422 "data_offset": 0, 00:10:40.422 "data_size": 65536 00:10:40.422 }, 00:10:40.422 { 00:10:40.422 "name": null, 00:10:40.422 "uuid": "e6876914-afda-4401-93b5-5a1e2448fe8f", 00:10:40.422 "is_configured": false, 00:10:40.422 "data_offset": 0, 00:10:40.422 "data_size": 65536 00:10:40.422 }, 00:10:40.422 { 00:10:40.422 "name": null, 00:10:40.422 "uuid": "2cf8b627-d211-4fdf-8173-4c8758dba086", 00:10:40.422 "is_configured": false, 00:10:40.422 "data_offset": 0, 00:10:40.422 "data_size": 65536 00:10:40.422 }, 00:10:40.422 { 00:10:40.422 "name": "BaseBdev4", 00:10:40.422 "uuid": "57ca4575-dbdd-4ac7-aba0-11c851287ee6", 00:10:40.422 "is_configured": true, 00:10:40.422 "data_offset": 0, 00:10:40.422 "data_size": 65536 00:10:40.422 } 00:10:40.422 ] 00:10:40.422 }' 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:40.422 19:00:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.682 [2024-12-05 19:00:58.154383] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:40.682 "name": "Existed_Raid", 00:10:40.682 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.682 "strip_size_kb": 0, 00:10:40.682 "state": "configuring", 00:10:40.682 "raid_level": "raid1", 00:10:40.682 "superblock": false, 00:10:40.682 "num_base_bdevs": 4, 00:10:40.682 "num_base_bdevs_discovered": 3, 00:10:40.682 "num_base_bdevs_operational": 4, 00:10:40.682 "base_bdevs_list": [ 00:10:40.682 { 00:10:40.682 "name": "BaseBdev1", 00:10:40.682 "uuid": "bb76236c-502f-4ccf-b213-c4886c9bcffe", 00:10:40.682 "is_configured": true, 00:10:40.682 "data_offset": 0, 00:10:40.682 "data_size": 65536 00:10:40.682 }, 00:10:40.682 { 00:10:40.682 "name": null, 00:10:40.682 "uuid": "e6876914-afda-4401-93b5-5a1e2448fe8f", 00:10:40.682 "is_configured": false, 00:10:40.682 "data_offset": 0, 00:10:40.682 "data_size": 65536 00:10:40.682 }, 00:10:40.682 { 00:10:40.682 "name": "BaseBdev3", 00:10:40.682 "uuid": "2cf8b627-d211-4fdf-8173-4c8758dba086", 00:10:40.682 "is_configured": true, 00:10:40.682 "data_offset": 0, 00:10:40.682 "data_size": 65536 00:10:40.682 }, 00:10:40.682 { 00:10:40.682 "name": "BaseBdev4", 00:10:40.682 "uuid": "57ca4575-dbdd-4ac7-aba0-11c851287ee6", 00:10:40.682 "is_configured": true, 00:10:40.682 "data_offset": 0, 00:10:40.682 "data_size": 65536 00:10:40.682 } 00:10:40.682 ] 00:10:40.682 }' 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:40.682 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.253 [2024-12-05 19:00:58.601692] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:41.253 "name": "Existed_Raid", 00:10:41.253 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:41.253 "strip_size_kb": 0, 00:10:41.253 "state": "configuring", 00:10:41.253 "raid_level": "raid1", 00:10:41.253 "superblock": false, 00:10:41.253 "num_base_bdevs": 4, 00:10:41.253 "num_base_bdevs_discovered": 2, 00:10:41.253 "num_base_bdevs_operational": 4, 00:10:41.253 "base_bdevs_list": [ 00:10:41.253 { 00:10:41.253 "name": null, 00:10:41.253 "uuid": "bb76236c-502f-4ccf-b213-c4886c9bcffe", 00:10:41.253 "is_configured": false, 00:10:41.253 "data_offset": 0, 00:10:41.253 "data_size": 65536 00:10:41.253 }, 00:10:41.253 { 00:10:41.253 "name": null, 00:10:41.253 "uuid": "e6876914-afda-4401-93b5-5a1e2448fe8f", 00:10:41.253 "is_configured": false, 00:10:41.253 "data_offset": 0, 00:10:41.253 "data_size": 65536 00:10:41.253 }, 00:10:41.253 { 00:10:41.253 "name": "BaseBdev3", 00:10:41.253 "uuid": "2cf8b627-d211-4fdf-8173-4c8758dba086", 00:10:41.253 "is_configured": true, 00:10:41.253 "data_offset": 0, 00:10:41.253 "data_size": 65536 00:10:41.253 }, 00:10:41.253 { 00:10:41.253 "name": "BaseBdev4", 00:10:41.253 "uuid": "57ca4575-dbdd-4ac7-aba0-11c851287ee6", 00:10:41.253 "is_configured": true, 00:10:41.253 "data_offset": 0, 00:10:41.253 "data_size": 65536 00:10:41.253 } 00:10:41.253 ] 00:10:41.253 }' 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:41.253 19:00:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.513 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.513 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.513 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.513 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:41.513 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.774 [2024-12-05 19:00:59.091280] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:41.774 "name": "Existed_Raid", 00:10:41.774 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:41.774 "strip_size_kb": 0, 00:10:41.774 "state": "configuring", 00:10:41.774 "raid_level": "raid1", 00:10:41.774 "superblock": false, 00:10:41.774 "num_base_bdevs": 4, 00:10:41.774 "num_base_bdevs_discovered": 3, 00:10:41.774 "num_base_bdevs_operational": 4, 00:10:41.774 "base_bdevs_list": [ 00:10:41.774 { 00:10:41.774 "name": null, 00:10:41.774 "uuid": "bb76236c-502f-4ccf-b213-c4886c9bcffe", 00:10:41.774 "is_configured": false, 00:10:41.774 "data_offset": 0, 00:10:41.774 "data_size": 65536 00:10:41.774 }, 00:10:41.774 { 00:10:41.774 "name": "BaseBdev2", 00:10:41.774 "uuid": "e6876914-afda-4401-93b5-5a1e2448fe8f", 00:10:41.774 "is_configured": true, 00:10:41.774 "data_offset": 0, 00:10:41.774 "data_size": 65536 00:10:41.774 }, 00:10:41.774 { 00:10:41.774 "name": "BaseBdev3", 00:10:41.774 "uuid": "2cf8b627-d211-4fdf-8173-4c8758dba086", 00:10:41.774 "is_configured": true, 00:10:41.774 "data_offset": 0, 00:10:41.774 "data_size": 65536 00:10:41.774 }, 00:10:41.774 { 00:10:41.774 "name": "BaseBdev4", 00:10:41.774 "uuid": "57ca4575-dbdd-4ac7-aba0-11c851287ee6", 00:10:41.774 "is_configured": true, 00:10:41.774 "data_offset": 0, 00:10:41.774 "data_size": 65536 00:10:41.774 } 00:10:41.774 ] 00:10:41.774 }' 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:41.774 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.034 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.034 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.034 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:42.034 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.034 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.034 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:42.034 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:42.034 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.034 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.034 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.034 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u bb76236c-502f-4ccf-b213-c4886c9bcffe 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.295 [2024-12-05 19:00:59.605440] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:42.295 [2024-12-05 19:00:59.605565] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:42.295 [2024-12-05 19:00:59.605596] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:42.295 [2024-12-05 19:00:59.605881] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:42.295 [2024-12-05 19:00:59.606069] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:42.295 [2024-12-05 19:00:59.606111] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:42.295 [2024-12-05 19:00:59.606324] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:42.295 NewBaseBdev 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@905 -- # local i 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.295 [ 00:10:42.295 { 00:10:42.295 "name": "NewBaseBdev", 00:10:42.295 "aliases": [ 00:10:42.295 "bb76236c-502f-4ccf-b213-c4886c9bcffe" 00:10:42.295 ], 00:10:42.295 "product_name": "Malloc disk", 00:10:42.295 "block_size": 512, 00:10:42.295 "num_blocks": 65536, 00:10:42.295 "uuid": "bb76236c-502f-4ccf-b213-c4886c9bcffe", 00:10:42.295 "assigned_rate_limits": { 00:10:42.295 "rw_ios_per_sec": 0, 00:10:42.295 "rw_mbytes_per_sec": 0, 00:10:42.295 "r_mbytes_per_sec": 0, 00:10:42.295 "w_mbytes_per_sec": 0 00:10:42.295 }, 00:10:42.295 "claimed": true, 00:10:42.295 "claim_type": "exclusive_write", 00:10:42.295 "zoned": false, 00:10:42.295 "supported_io_types": { 00:10:42.295 "read": true, 00:10:42.295 "write": true, 00:10:42.295 "unmap": true, 00:10:42.295 "flush": true, 00:10:42.295 "reset": true, 00:10:42.295 "nvme_admin": false, 00:10:42.295 "nvme_io": false, 00:10:42.295 "nvme_io_md": false, 00:10:42.295 "write_zeroes": true, 00:10:42.295 "zcopy": true, 00:10:42.295 "get_zone_info": false, 00:10:42.295 "zone_management": false, 00:10:42.295 "zone_append": false, 00:10:42.295 "compare": false, 00:10:42.295 "compare_and_write": false, 00:10:42.295 "abort": true, 00:10:42.295 "seek_hole": false, 00:10:42.295 "seek_data": false, 00:10:42.295 "copy": true, 00:10:42.295 "nvme_iov_md": false 00:10:42.295 }, 00:10:42.295 "memory_domains": [ 00:10:42.295 { 00:10:42.295 "dma_device_id": "system", 00:10:42.295 "dma_device_type": 1 00:10:42.295 }, 00:10:42.295 { 00:10:42.295 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.295 "dma_device_type": 2 00:10:42.295 } 00:10:42.295 ], 00:10:42.295 "driver_specific": {} 00:10:42.295 } 00:10:42.295 ] 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:42.295 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:42.296 "name": "Existed_Raid", 00:10:42.296 "uuid": "a124455a-10ed-4f56-b079-e02559df409b", 00:10:42.296 "strip_size_kb": 0, 00:10:42.296 "state": "online", 00:10:42.296 "raid_level": "raid1", 00:10:42.296 "superblock": false, 00:10:42.296 "num_base_bdevs": 4, 00:10:42.296 "num_base_bdevs_discovered": 4, 00:10:42.296 "num_base_bdevs_operational": 4, 00:10:42.296 "base_bdevs_list": [ 00:10:42.296 { 00:10:42.296 "name": "NewBaseBdev", 00:10:42.296 "uuid": "bb76236c-502f-4ccf-b213-c4886c9bcffe", 00:10:42.296 "is_configured": true, 00:10:42.296 "data_offset": 0, 00:10:42.296 "data_size": 65536 00:10:42.296 }, 00:10:42.296 { 00:10:42.296 "name": "BaseBdev2", 00:10:42.296 "uuid": "e6876914-afda-4401-93b5-5a1e2448fe8f", 00:10:42.296 "is_configured": true, 00:10:42.296 "data_offset": 0, 00:10:42.296 "data_size": 65536 00:10:42.296 }, 00:10:42.296 { 00:10:42.296 "name": "BaseBdev3", 00:10:42.296 "uuid": "2cf8b627-d211-4fdf-8173-4c8758dba086", 00:10:42.296 "is_configured": true, 00:10:42.296 "data_offset": 0, 00:10:42.296 "data_size": 65536 00:10:42.296 }, 00:10:42.296 { 00:10:42.296 "name": "BaseBdev4", 00:10:42.296 "uuid": "57ca4575-dbdd-4ac7-aba0-11c851287ee6", 00:10:42.296 "is_configured": true, 00:10:42.296 "data_offset": 0, 00:10:42.296 "data_size": 65536 00:10:42.296 } 00:10:42.296 ] 00:10:42.296 }' 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:42.296 19:00:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.556 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:42.556 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:42.556 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:42.556 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:42.556 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:42.556 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:42.556 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:42.556 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:42.556 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.556 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.556 [2024-12-05 19:01:00.069044] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:42.556 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.556 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:42.556 "name": "Existed_Raid", 00:10:42.556 "aliases": [ 00:10:42.556 "a124455a-10ed-4f56-b079-e02559df409b" 00:10:42.556 ], 00:10:42.556 "product_name": "Raid Volume", 00:10:42.556 "block_size": 512, 00:10:42.556 "num_blocks": 65536, 00:10:42.556 "uuid": "a124455a-10ed-4f56-b079-e02559df409b", 00:10:42.556 "assigned_rate_limits": { 00:10:42.556 "rw_ios_per_sec": 0, 00:10:42.556 "rw_mbytes_per_sec": 0, 00:10:42.556 "r_mbytes_per_sec": 0, 00:10:42.556 "w_mbytes_per_sec": 0 00:10:42.556 }, 00:10:42.556 "claimed": false, 00:10:42.556 "zoned": false, 00:10:42.556 "supported_io_types": { 00:10:42.556 "read": true, 00:10:42.556 "write": true, 00:10:42.556 "unmap": false, 00:10:42.556 "flush": false, 00:10:42.556 "reset": true, 00:10:42.556 "nvme_admin": false, 00:10:42.556 "nvme_io": false, 00:10:42.556 "nvme_io_md": false, 00:10:42.556 "write_zeroes": true, 00:10:42.556 "zcopy": false, 00:10:42.556 "get_zone_info": false, 00:10:42.556 "zone_management": false, 00:10:42.556 "zone_append": false, 00:10:42.556 "compare": false, 00:10:42.556 "compare_and_write": false, 00:10:42.556 "abort": false, 00:10:42.556 "seek_hole": false, 00:10:42.556 "seek_data": false, 00:10:42.556 "copy": false, 00:10:42.556 "nvme_iov_md": false 00:10:42.556 }, 00:10:42.556 "memory_domains": [ 00:10:42.556 { 00:10:42.556 "dma_device_id": "system", 00:10:42.556 "dma_device_type": 1 00:10:42.556 }, 00:10:42.556 { 00:10:42.556 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.556 "dma_device_type": 2 00:10:42.556 }, 00:10:42.556 { 00:10:42.556 "dma_device_id": "system", 00:10:42.556 "dma_device_type": 1 00:10:42.556 }, 00:10:42.556 { 00:10:42.556 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.556 "dma_device_type": 2 00:10:42.556 }, 00:10:42.557 { 00:10:42.557 "dma_device_id": "system", 00:10:42.557 "dma_device_type": 1 00:10:42.557 }, 00:10:42.557 { 00:10:42.557 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.557 "dma_device_type": 2 00:10:42.557 }, 00:10:42.557 { 00:10:42.557 "dma_device_id": "system", 00:10:42.557 "dma_device_type": 1 00:10:42.557 }, 00:10:42.557 { 00:10:42.557 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.557 "dma_device_type": 2 00:10:42.557 } 00:10:42.557 ], 00:10:42.557 "driver_specific": { 00:10:42.557 "raid": { 00:10:42.557 "uuid": "a124455a-10ed-4f56-b079-e02559df409b", 00:10:42.557 "strip_size_kb": 0, 00:10:42.557 "state": "online", 00:10:42.557 "raid_level": "raid1", 00:10:42.557 "superblock": false, 00:10:42.557 "num_base_bdevs": 4, 00:10:42.557 "num_base_bdevs_discovered": 4, 00:10:42.557 "num_base_bdevs_operational": 4, 00:10:42.557 "base_bdevs_list": [ 00:10:42.557 { 00:10:42.557 "name": "NewBaseBdev", 00:10:42.557 "uuid": "bb76236c-502f-4ccf-b213-c4886c9bcffe", 00:10:42.557 "is_configured": true, 00:10:42.557 "data_offset": 0, 00:10:42.557 "data_size": 65536 00:10:42.557 }, 00:10:42.557 { 00:10:42.557 "name": "BaseBdev2", 00:10:42.557 "uuid": "e6876914-afda-4401-93b5-5a1e2448fe8f", 00:10:42.557 "is_configured": true, 00:10:42.557 "data_offset": 0, 00:10:42.557 "data_size": 65536 00:10:42.557 }, 00:10:42.557 { 00:10:42.557 "name": "BaseBdev3", 00:10:42.557 "uuid": "2cf8b627-d211-4fdf-8173-4c8758dba086", 00:10:42.557 "is_configured": true, 00:10:42.557 "data_offset": 0, 00:10:42.557 "data_size": 65536 00:10:42.557 }, 00:10:42.557 { 00:10:42.557 "name": "BaseBdev4", 00:10:42.557 "uuid": "57ca4575-dbdd-4ac7-aba0-11c851287ee6", 00:10:42.557 "is_configured": true, 00:10:42.557 "data_offset": 0, 00:10:42.557 "data_size": 65536 00:10:42.557 } 00:10:42.557 ] 00:10:42.557 } 00:10:42.557 } 00:10:42.557 }' 00:10:42.557 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:42.817 BaseBdev2 00:10:42.817 BaseBdev3 00:10:42.817 BaseBdev4' 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.817 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:43.077 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:43.077 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:43.077 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:43.077 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:43.077 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.077 [2024-12-05 19:01:00.380192] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:43.077 [2024-12-05 19:01:00.380257] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:43.077 [2024-12-05 19:01:00.380355] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:43.077 [2024-12-05 19:01:00.380615] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:43.077 [2024-12-05 19:01:00.380678] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:43.077 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:43.077 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 82931 00:10:43.077 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 82931 ']' 00:10:43.077 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@958 -- # kill -0 82931 00:10:43.077 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # uname 00:10:43.077 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:43.077 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 82931 00:10:43.077 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:43.078 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:43.078 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 82931' 00:10:43.078 killing process with pid 82931 00:10:43.078 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@973 -- # kill 82931 00:10:43.078 [2024-12-05 19:01:00.416783] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:43.078 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@978 -- # wait 82931 00:10:43.078 [2024-12-05 19:01:00.456106] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:43.338 19:01:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:43.339 ************************************ 00:10:43.339 END TEST raid_state_function_test 00:10:43.339 ************************************ 00:10:43.339 00:10:43.339 real 0m9.274s 00:10:43.339 user 0m15.915s 00:10:43.339 sys 0m1.845s 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.339 19:01:00 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 4 true 00:10:43.339 19:01:00 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:43.339 19:01:00 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:43.339 19:01:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:43.339 ************************************ 00:10:43.339 START TEST raid_state_function_test_sb 00:10:43.339 ************************************ 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 4 true 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:43.339 Process raid pid: 83521 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=83521 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 83521' 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 83521 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 83521 ']' 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:43.339 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:43.339 19:01:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.339 [2024-12-05 19:01:00.831427] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:10:43.339 [2024-12-05 19:01:00.831645] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:43.599 [2024-12-05 19:01:00.985866] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:43.600 [2024-12-05 19:01:01.010642] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:43.600 [2024-12-05 19:01:01.053572] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:43.600 [2024-12-05 19:01:01.053753] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.169 [2024-12-05 19:01:01.660693] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:44.169 [2024-12-05 19:01:01.660810] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:44.169 [2024-12-05 19:01:01.660843] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:44.169 [2024-12-05 19:01:01.660866] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:44.169 [2024-12-05 19:01:01.660884] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:44.169 [2024-12-05 19:01:01.660908] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:44.169 [2024-12-05 19:01:01.660926] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:44.169 [2024-12-05 19:01:01.660947] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:44.169 "name": "Existed_Raid", 00:10:44.169 "uuid": "515aa677-7752-43ac-acf5-b98748108140", 00:10:44.169 "strip_size_kb": 0, 00:10:44.169 "state": "configuring", 00:10:44.169 "raid_level": "raid1", 00:10:44.169 "superblock": true, 00:10:44.169 "num_base_bdevs": 4, 00:10:44.169 "num_base_bdevs_discovered": 0, 00:10:44.169 "num_base_bdevs_operational": 4, 00:10:44.169 "base_bdevs_list": [ 00:10:44.169 { 00:10:44.169 "name": "BaseBdev1", 00:10:44.169 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:44.169 "is_configured": false, 00:10:44.169 "data_offset": 0, 00:10:44.169 "data_size": 0 00:10:44.169 }, 00:10:44.169 { 00:10:44.169 "name": "BaseBdev2", 00:10:44.169 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:44.169 "is_configured": false, 00:10:44.169 "data_offset": 0, 00:10:44.169 "data_size": 0 00:10:44.169 }, 00:10:44.169 { 00:10:44.169 "name": "BaseBdev3", 00:10:44.169 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:44.169 "is_configured": false, 00:10:44.169 "data_offset": 0, 00:10:44.169 "data_size": 0 00:10:44.169 }, 00:10:44.169 { 00:10:44.169 "name": "BaseBdev4", 00:10:44.169 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:44.169 "is_configured": false, 00:10:44.169 "data_offset": 0, 00:10:44.169 "data_size": 0 00:10:44.169 } 00:10:44.169 ] 00:10:44.169 }' 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:44.169 19:01:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.805 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:44.805 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.805 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.805 [2024-12-05 19:01:02.071892] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:44.805 [2024-12-05 19:01:02.071971] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:44.805 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.805 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.806 [2024-12-05 19:01:02.083891] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:44.806 [2024-12-05 19:01:02.083986] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:44.806 [2024-12-05 19:01:02.084013] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:44.806 [2024-12-05 19:01:02.084035] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:44.806 [2024-12-05 19:01:02.084053] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:44.806 [2024-12-05 19:01:02.084072] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:44.806 [2024-12-05 19:01:02.084089] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:44.806 [2024-12-05 19:01:02.084109] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.806 [2024-12-05 19:01:02.104740] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:44.806 BaseBdev1 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.806 [ 00:10:44.806 { 00:10:44.806 "name": "BaseBdev1", 00:10:44.806 "aliases": [ 00:10:44.806 "cd810a24-4308-4970-9583-787490c8e1b0" 00:10:44.806 ], 00:10:44.806 "product_name": "Malloc disk", 00:10:44.806 "block_size": 512, 00:10:44.806 "num_blocks": 65536, 00:10:44.806 "uuid": "cd810a24-4308-4970-9583-787490c8e1b0", 00:10:44.806 "assigned_rate_limits": { 00:10:44.806 "rw_ios_per_sec": 0, 00:10:44.806 "rw_mbytes_per_sec": 0, 00:10:44.806 "r_mbytes_per_sec": 0, 00:10:44.806 "w_mbytes_per_sec": 0 00:10:44.806 }, 00:10:44.806 "claimed": true, 00:10:44.806 "claim_type": "exclusive_write", 00:10:44.806 "zoned": false, 00:10:44.806 "supported_io_types": { 00:10:44.806 "read": true, 00:10:44.806 "write": true, 00:10:44.806 "unmap": true, 00:10:44.806 "flush": true, 00:10:44.806 "reset": true, 00:10:44.806 "nvme_admin": false, 00:10:44.806 "nvme_io": false, 00:10:44.806 "nvme_io_md": false, 00:10:44.806 "write_zeroes": true, 00:10:44.806 "zcopy": true, 00:10:44.806 "get_zone_info": false, 00:10:44.806 "zone_management": false, 00:10:44.806 "zone_append": false, 00:10:44.806 "compare": false, 00:10:44.806 "compare_and_write": false, 00:10:44.806 "abort": true, 00:10:44.806 "seek_hole": false, 00:10:44.806 "seek_data": false, 00:10:44.806 "copy": true, 00:10:44.806 "nvme_iov_md": false 00:10:44.806 }, 00:10:44.806 "memory_domains": [ 00:10:44.806 { 00:10:44.806 "dma_device_id": "system", 00:10:44.806 "dma_device_type": 1 00:10:44.806 }, 00:10:44.806 { 00:10:44.806 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:44.806 "dma_device_type": 2 00:10:44.806 } 00:10:44.806 ], 00:10:44.806 "driver_specific": {} 00:10:44.806 } 00:10:44.806 ] 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:44.806 "name": "Existed_Raid", 00:10:44.806 "uuid": "8226cb88-f907-4759-b2e2-18d8bffd6889", 00:10:44.806 "strip_size_kb": 0, 00:10:44.806 "state": "configuring", 00:10:44.806 "raid_level": "raid1", 00:10:44.806 "superblock": true, 00:10:44.806 "num_base_bdevs": 4, 00:10:44.806 "num_base_bdevs_discovered": 1, 00:10:44.806 "num_base_bdevs_operational": 4, 00:10:44.806 "base_bdevs_list": [ 00:10:44.806 { 00:10:44.806 "name": "BaseBdev1", 00:10:44.806 "uuid": "cd810a24-4308-4970-9583-787490c8e1b0", 00:10:44.806 "is_configured": true, 00:10:44.806 "data_offset": 2048, 00:10:44.806 "data_size": 63488 00:10:44.806 }, 00:10:44.806 { 00:10:44.806 "name": "BaseBdev2", 00:10:44.806 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:44.806 "is_configured": false, 00:10:44.806 "data_offset": 0, 00:10:44.806 "data_size": 0 00:10:44.806 }, 00:10:44.806 { 00:10:44.806 "name": "BaseBdev3", 00:10:44.806 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:44.806 "is_configured": false, 00:10:44.806 "data_offset": 0, 00:10:44.806 "data_size": 0 00:10:44.806 }, 00:10:44.806 { 00:10:44.806 "name": "BaseBdev4", 00:10:44.806 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:44.806 "is_configured": false, 00:10:44.806 "data_offset": 0, 00:10:44.806 "data_size": 0 00:10:44.806 } 00:10:44.806 ] 00:10:44.806 }' 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:44.806 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.083 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:45.083 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.083 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.083 [2024-12-05 19:01:02.567952] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:45.083 [2024-12-05 19:01:02.568032] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:45.083 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.083 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:45.083 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.083 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.083 [2024-12-05 19:01:02.579965] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:45.083 [2024-12-05 19:01:02.581826] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:45.083 [2024-12-05 19:01:02.581913] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:45.083 [2024-12-05 19:01:02.581939] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:45.083 [2024-12-05 19:01:02.581961] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:45.083 [2024-12-05 19:01:02.581978] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:45.083 [2024-12-05 19:01:02.581996] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:45.083 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.083 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:45.083 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:45.083 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:45.083 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:45.084 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:45.084 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:45.084 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:45.084 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:45.084 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:45.084 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:45.084 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:45.084 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:45.084 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.084 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:45.084 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.084 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.084 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.355 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:45.355 "name": "Existed_Raid", 00:10:45.355 "uuid": "23ed8fe5-5c6e-45df-a9de-cb82c9b30d5f", 00:10:45.355 "strip_size_kb": 0, 00:10:45.355 "state": "configuring", 00:10:45.355 "raid_level": "raid1", 00:10:45.355 "superblock": true, 00:10:45.355 "num_base_bdevs": 4, 00:10:45.355 "num_base_bdevs_discovered": 1, 00:10:45.355 "num_base_bdevs_operational": 4, 00:10:45.355 "base_bdevs_list": [ 00:10:45.355 { 00:10:45.355 "name": "BaseBdev1", 00:10:45.355 "uuid": "cd810a24-4308-4970-9583-787490c8e1b0", 00:10:45.355 "is_configured": true, 00:10:45.355 "data_offset": 2048, 00:10:45.355 "data_size": 63488 00:10:45.355 }, 00:10:45.355 { 00:10:45.355 "name": "BaseBdev2", 00:10:45.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:45.355 "is_configured": false, 00:10:45.355 "data_offset": 0, 00:10:45.355 "data_size": 0 00:10:45.355 }, 00:10:45.355 { 00:10:45.355 "name": "BaseBdev3", 00:10:45.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:45.355 "is_configured": false, 00:10:45.355 "data_offset": 0, 00:10:45.355 "data_size": 0 00:10:45.355 }, 00:10:45.355 { 00:10:45.355 "name": "BaseBdev4", 00:10:45.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:45.355 "is_configured": false, 00:10:45.355 "data_offset": 0, 00:10:45.355 "data_size": 0 00:10:45.355 } 00:10:45.355 ] 00:10:45.355 }' 00:10:45.355 19:01:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:45.355 19:01:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.614 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:45.614 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.614 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.614 [2024-12-05 19:01:03.046181] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:45.614 BaseBdev2 00:10:45.614 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.615 [ 00:10:45.615 { 00:10:45.615 "name": "BaseBdev2", 00:10:45.615 "aliases": [ 00:10:45.615 "9cf7245e-b1af-417d-98bd-539129a83e4b" 00:10:45.615 ], 00:10:45.615 "product_name": "Malloc disk", 00:10:45.615 "block_size": 512, 00:10:45.615 "num_blocks": 65536, 00:10:45.615 "uuid": "9cf7245e-b1af-417d-98bd-539129a83e4b", 00:10:45.615 "assigned_rate_limits": { 00:10:45.615 "rw_ios_per_sec": 0, 00:10:45.615 "rw_mbytes_per_sec": 0, 00:10:45.615 "r_mbytes_per_sec": 0, 00:10:45.615 "w_mbytes_per_sec": 0 00:10:45.615 }, 00:10:45.615 "claimed": true, 00:10:45.615 "claim_type": "exclusive_write", 00:10:45.615 "zoned": false, 00:10:45.615 "supported_io_types": { 00:10:45.615 "read": true, 00:10:45.615 "write": true, 00:10:45.615 "unmap": true, 00:10:45.615 "flush": true, 00:10:45.615 "reset": true, 00:10:45.615 "nvme_admin": false, 00:10:45.615 "nvme_io": false, 00:10:45.615 "nvme_io_md": false, 00:10:45.615 "write_zeroes": true, 00:10:45.615 "zcopy": true, 00:10:45.615 "get_zone_info": false, 00:10:45.615 "zone_management": false, 00:10:45.615 "zone_append": false, 00:10:45.615 "compare": false, 00:10:45.615 "compare_and_write": false, 00:10:45.615 "abort": true, 00:10:45.615 "seek_hole": false, 00:10:45.615 "seek_data": false, 00:10:45.615 "copy": true, 00:10:45.615 "nvme_iov_md": false 00:10:45.615 }, 00:10:45.615 "memory_domains": [ 00:10:45.615 { 00:10:45.615 "dma_device_id": "system", 00:10:45.615 "dma_device_type": 1 00:10:45.615 }, 00:10:45.615 { 00:10:45.615 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:45.615 "dma_device_type": 2 00:10:45.615 } 00:10:45.615 ], 00:10:45.615 "driver_specific": {} 00:10:45.615 } 00:10:45.615 ] 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:45.615 "name": "Existed_Raid", 00:10:45.615 "uuid": "23ed8fe5-5c6e-45df-a9de-cb82c9b30d5f", 00:10:45.615 "strip_size_kb": 0, 00:10:45.615 "state": "configuring", 00:10:45.615 "raid_level": "raid1", 00:10:45.615 "superblock": true, 00:10:45.615 "num_base_bdevs": 4, 00:10:45.615 "num_base_bdevs_discovered": 2, 00:10:45.615 "num_base_bdevs_operational": 4, 00:10:45.615 "base_bdevs_list": [ 00:10:45.615 { 00:10:45.615 "name": "BaseBdev1", 00:10:45.615 "uuid": "cd810a24-4308-4970-9583-787490c8e1b0", 00:10:45.615 "is_configured": true, 00:10:45.615 "data_offset": 2048, 00:10:45.615 "data_size": 63488 00:10:45.615 }, 00:10:45.615 { 00:10:45.615 "name": "BaseBdev2", 00:10:45.615 "uuid": "9cf7245e-b1af-417d-98bd-539129a83e4b", 00:10:45.615 "is_configured": true, 00:10:45.615 "data_offset": 2048, 00:10:45.615 "data_size": 63488 00:10:45.615 }, 00:10:45.615 { 00:10:45.615 "name": "BaseBdev3", 00:10:45.615 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:45.615 "is_configured": false, 00:10:45.615 "data_offset": 0, 00:10:45.615 "data_size": 0 00:10:45.615 }, 00:10:45.615 { 00:10:45.615 "name": "BaseBdev4", 00:10:45.615 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:45.615 "is_configured": false, 00:10:45.615 "data_offset": 0, 00:10:45.615 "data_size": 0 00:10:45.615 } 00:10:45.615 ] 00:10:45.615 }' 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:45.615 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.184 [2024-12-05 19:01:03.542955] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:46.184 BaseBdev3 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.184 [ 00:10:46.184 { 00:10:46.184 "name": "BaseBdev3", 00:10:46.184 "aliases": [ 00:10:46.184 "efc36ed2-a2be-4d31-9cd9-2a77b4c31e8b" 00:10:46.184 ], 00:10:46.184 "product_name": "Malloc disk", 00:10:46.184 "block_size": 512, 00:10:46.184 "num_blocks": 65536, 00:10:46.184 "uuid": "efc36ed2-a2be-4d31-9cd9-2a77b4c31e8b", 00:10:46.184 "assigned_rate_limits": { 00:10:46.184 "rw_ios_per_sec": 0, 00:10:46.184 "rw_mbytes_per_sec": 0, 00:10:46.184 "r_mbytes_per_sec": 0, 00:10:46.184 "w_mbytes_per_sec": 0 00:10:46.184 }, 00:10:46.184 "claimed": true, 00:10:46.184 "claim_type": "exclusive_write", 00:10:46.184 "zoned": false, 00:10:46.184 "supported_io_types": { 00:10:46.184 "read": true, 00:10:46.184 "write": true, 00:10:46.184 "unmap": true, 00:10:46.184 "flush": true, 00:10:46.184 "reset": true, 00:10:46.184 "nvme_admin": false, 00:10:46.184 "nvme_io": false, 00:10:46.184 "nvme_io_md": false, 00:10:46.184 "write_zeroes": true, 00:10:46.184 "zcopy": true, 00:10:46.184 "get_zone_info": false, 00:10:46.184 "zone_management": false, 00:10:46.184 "zone_append": false, 00:10:46.184 "compare": false, 00:10:46.184 "compare_and_write": false, 00:10:46.184 "abort": true, 00:10:46.184 "seek_hole": false, 00:10:46.184 "seek_data": false, 00:10:46.184 "copy": true, 00:10:46.184 "nvme_iov_md": false 00:10:46.184 }, 00:10:46.184 "memory_domains": [ 00:10:46.184 { 00:10:46.184 "dma_device_id": "system", 00:10:46.184 "dma_device_type": 1 00:10:46.184 }, 00:10:46.184 { 00:10:46.184 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:46.184 "dma_device_type": 2 00:10:46.184 } 00:10:46.184 ], 00:10:46.184 "driver_specific": {} 00:10:46.184 } 00:10:46.184 ] 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.184 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.185 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.185 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:46.185 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.185 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.185 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.185 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.185 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:46.185 "name": "Existed_Raid", 00:10:46.185 "uuid": "23ed8fe5-5c6e-45df-a9de-cb82c9b30d5f", 00:10:46.185 "strip_size_kb": 0, 00:10:46.185 "state": "configuring", 00:10:46.185 "raid_level": "raid1", 00:10:46.185 "superblock": true, 00:10:46.185 "num_base_bdevs": 4, 00:10:46.185 "num_base_bdevs_discovered": 3, 00:10:46.185 "num_base_bdevs_operational": 4, 00:10:46.185 "base_bdevs_list": [ 00:10:46.185 { 00:10:46.185 "name": "BaseBdev1", 00:10:46.185 "uuid": "cd810a24-4308-4970-9583-787490c8e1b0", 00:10:46.185 "is_configured": true, 00:10:46.185 "data_offset": 2048, 00:10:46.185 "data_size": 63488 00:10:46.185 }, 00:10:46.185 { 00:10:46.185 "name": "BaseBdev2", 00:10:46.185 "uuid": "9cf7245e-b1af-417d-98bd-539129a83e4b", 00:10:46.185 "is_configured": true, 00:10:46.185 "data_offset": 2048, 00:10:46.185 "data_size": 63488 00:10:46.185 }, 00:10:46.185 { 00:10:46.185 "name": "BaseBdev3", 00:10:46.185 "uuid": "efc36ed2-a2be-4d31-9cd9-2a77b4c31e8b", 00:10:46.185 "is_configured": true, 00:10:46.185 "data_offset": 2048, 00:10:46.185 "data_size": 63488 00:10:46.185 }, 00:10:46.185 { 00:10:46.185 "name": "BaseBdev4", 00:10:46.185 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:46.185 "is_configured": false, 00:10:46.185 "data_offset": 0, 00:10:46.185 "data_size": 0 00:10:46.185 } 00:10:46.185 ] 00:10:46.185 }' 00:10:46.185 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:46.185 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.444 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:46.444 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.444 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.444 [2024-12-05 19:01:03.993207] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:46.444 [2024-12-05 19:01:03.993503] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:46.444 [2024-12-05 19:01:03.993559] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:46.444 [2024-12-05 19:01:03.993854] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:46.444 BaseBdev4 00:10:46.444 [2024-12-05 19:01:03.994059] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:46.444 [2024-12-05 19:01:03.994079] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:46.444 [2024-12-05 19:01:03.994213] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:46.444 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.444 19:01:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:46.444 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:46.444 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:46.444 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:46.444 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:46.444 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:46.444 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:46.444 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.444 19:01:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.704 [ 00:10:46.704 { 00:10:46.704 "name": "BaseBdev4", 00:10:46.704 "aliases": [ 00:10:46.704 "6a69c794-5070-420e-9852-bbd207e6c6d7" 00:10:46.704 ], 00:10:46.704 "product_name": "Malloc disk", 00:10:46.704 "block_size": 512, 00:10:46.704 "num_blocks": 65536, 00:10:46.704 "uuid": "6a69c794-5070-420e-9852-bbd207e6c6d7", 00:10:46.704 "assigned_rate_limits": { 00:10:46.704 "rw_ios_per_sec": 0, 00:10:46.704 "rw_mbytes_per_sec": 0, 00:10:46.704 "r_mbytes_per_sec": 0, 00:10:46.704 "w_mbytes_per_sec": 0 00:10:46.704 }, 00:10:46.704 "claimed": true, 00:10:46.704 "claim_type": "exclusive_write", 00:10:46.704 "zoned": false, 00:10:46.704 "supported_io_types": { 00:10:46.704 "read": true, 00:10:46.704 "write": true, 00:10:46.704 "unmap": true, 00:10:46.704 "flush": true, 00:10:46.704 "reset": true, 00:10:46.704 "nvme_admin": false, 00:10:46.704 "nvme_io": false, 00:10:46.704 "nvme_io_md": false, 00:10:46.704 "write_zeroes": true, 00:10:46.704 "zcopy": true, 00:10:46.704 "get_zone_info": false, 00:10:46.704 "zone_management": false, 00:10:46.704 "zone_append": false, 00:10:46.704 "compare": false, 00:10:46.704 "compare_and_write": false, 00:10:46.704 "abort": true, 00:10:46.704 "seek_hole": false, 00:10:46.704 "seek_data": false, 00:10:46.704 "copy": true, 00:10:46.704 "nvme_iov_md": false 00:10:46.704 }, 00:10:46.704 "memory_domains": [ 00:10:46.704 { 00:10:46.704 "dma_device_id": "system", 00:10:46.704 "dma_device_type": 1 00:10:46.704 }, 00:10:46.704 { 00:10:46.704 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:46.704 "dma_device_type": 2 00:10:46.704 } 00:10:46.704 ], 00:10:46.704 "driver_specific": {} 00:10:46.704 } 00:10:46.704 ] 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.704 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:46.704 "name": "Existed_Raid", 00:10:46.704 "uuid": "23ed8fe5-5c6e-45df-a9de-cb82c9b30d5f", 00:10:46.704 "strip_size_kb": 0, 00:10:46.704 "state": "online", 00:10:46.704 "raid_level": "raid1", 00:10:46.704 "superblock": true, 00:10:46.704 "num_base_bdevs": 4, 00:10:46.704 "num_base_bdevs_discovered": 4, 00:10:46.704 "num_base_bdevs_operational": 4, 00:10:46.704 "base_bdevs_list": [ 00:10:46.704 { 00:10:46.704 "name": "BaseBdev1", 00:10:46.704 "uuid": "cd810a24-4308-4970-9583-787490c8e1b0", 00:10:46.704 "is_configured": true, 00:10:46.704 "data_offset": 2048, 00:10:46.704 "data_size": 63488 00:10:46.704 }, 00:10:46.704 { 00:10:46.704 "name": "BaseBdev2", 00:10:46.704 "uuid": "9cf7245e-b1af-417d-98bd-539129a83e4b", 00:10:46.704 "is_configured": true, 00:10:46.704 "data_offset": 2048, 00:10:46.704 "data_size": 63488 00:10:46.704 }, 00:10:46.704 { 00:10:46.704 "name": "BaseBdev3", 00:10:46.704 "uuid": "efc36ed2-a2be-4d31-9cd9-2a77b4c31e8b", 00:10:46.704 "is_configured": true, 00:10:46.704 "data_offset": 2048, 00:10:46.704 "data_size": 63488 00:10:46.704 }, 00:10:46.705 { 00:10:46.705 "name": "BaseBdev4", 00:10:46.705 "uuid": "6a69c794-5070-420e-9852-bbd207e6c6d7", 00:10:46.705 "is_configured": true, 00:10:46.705 "data_offset": 2048, 00:10:46.705 "data_size": 63488 00:10:46.705 } 00:10:46.705 ] 00:10:46.705 }' 00:10:46.705 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:46.705 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.965 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:46.965 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:46.965 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:46.965 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:46.965 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:46.965 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:46.965 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:46.965 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:46.965 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:46.965 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.965 [2024-12-05 19:01:04.448884] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:46.965 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:46.965 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:46.965 "name": "Existed_Raid", 00:10:46.965 "aliases": [ 00:10:46.965 "23ed8fe5-5c6e-45df-a9de-cb82c9b30d5f" 00:10:46.965 ], 00:10:46.965 "product_name": "Raid Volume", 00:10:46.965 "block_size": 512, 00:10:46.965 "num_blocks": 63488, 00:10:46.965 "uuid": "23ed8fe5-5c6e-45df-a9de-cb82c9b30d5f", 00:10:46.965 "assigned_rate_limits": { 00:10:46.965 "rw_ios_per_sec": 0, 00:10:46.965 "rw_mbytes_per_sec": 0, 00:10:46.965 "r_mbytes_per_sec": 0, 00:10:46.965 "w_mbytes_per_sec": 0 00:10:46.965 }, 00:10:46.965 "claimed": false, 00:10:46.965 "zoned": false, 00:10:46.965 "supported_io_types": { 00:10:46.965 "read": true, 00:10:46.965 "write": true, 00:10:46.965 "unmap": false, 00:10:46.965 "flush": false, 00:10:46.965 "reset": true, 00:10:46.965 "nvme_admin": false, 00:10:46.965 "nvme_io": false, 00:10:46.965 "nvme_io_md": false, 00:10:46.965 "write_zeroes": true, 00:10:46.965 "zcopy": false, 00:10:46.965 "get_zone_info": false, 00:10:46.965 "zone_management": false, 00:10:46.965 "zone_append": false, 00:10:46.965 "compare": false, 00:10:46.965 "compare_and_write": false, 00:10:46.965 "abort": false, 00:10:46.965 "seek_hole": false, 00:10:46.965 "seek_data": false, 00:10:46.965 "copy": false, 00:10:46.965 "nvme_iov_md": false 00:10:46.965 }, 00:10:46.965 "memory_domains": [ 00:10:46.965 { 00:10:46.965 "dma_device_id": "system", 00:10:46.965 "dma_device_type": 1 00:10:46.965 }, 00:10:46.965 { 00:10:46.965 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:46.965 "dma_device_type": 2 00:10:46.965 }, 00:10:46.965 { 00:10:46.965 "dma_device_id": "system", 00:10:46.965 "dma_device_type": 1 00:10:46.965 }, 00:10:46.965 { 00:10:46.965 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:46.965 "dma_device_type": 2 00:10:46.965 }, 00:10:46.965 { 00:10:46.965 "dma_device_id": "system", 00:10:46.965 "dma_device_type": 1 00:10:46.965 }, 00:10:46.965 { 00:10:46.965 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:46.965 "dma_device_type": 2 00:10:46.965 }, 00:10:46.965 { 00:10:46.965 "dma_device_id": "system", 00:10:46.965 "dma_device_type": 1 00:10:46.965 }, 00:10:46.965 { 00:10:46.965 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:46.965 "dma_device_type": 2 00:10:46.965 } 00:10:46.965 ], 00:10:46.965 "driver_specific": { 00:10:46.965 "raid": { 00:10:46.965 "uuid": "23ed8fe5-5c6e-45df-a9de-cb82c9b30d5f", 00:10:46.965 "strip_size_kb": 0, 00:10:46.965 "state": "online", 00:10:46.965 "raid_level": "raid1", 00:10:46.965 "superblock": true, 00:10:46.965 "num_base_bdevs": 4, 00:10:46.965 "num_base_bdevs_discovered": 4, 00:10:46.965 "num_base_bdevs_operational": 4, 00:10:46.965 "base_bdevs_list": [ 00:10:46.965 { 00:10:46.965 "name": "BaseBdev1", 00:10:46.965 "uuid": "cd810a24-4308-4970-9583-787490c8e1b0", 00:10:46.965 "is_configured": true, 00:10:46.965 "data_offset": 2048, 00:10:46.965 "data_size": 63488 00:10:46.965 }, 00:10:46.965 { 00:10:46.965 "name": "BaseBdev2", 00:10:46.965 "uuid": "9cf7245e-b1af-417d-98bd-539129a83e4b", 00:10:46.965 "is_configured": true, 00:10:46.965 "data_offset": 2048, 00:10:46.965 "data_size": 63488 00:10:46.965 }, 00:10:46.965 { 00:10:46.965 "name": "BaseBdev3", 00:10:46.965 "uuid": "efc36ed2-a2be-4d31-9cd9-2a77b4c31e8b", 00:10:46.965 "is_configured": true, 00:10:46.965 "data_offset": 2048, 00:10:46.965 "data_size": 63488 00:10:46.965 }, 00:10:46.965 { 00:10:46.965 "name": "BaseBdev4", 00:10:46.965 "uuid": "6a69c794-5070-420e-9852-bbd207e6c6d7", 00:10:46.965 "is_configured": true, 00:10:46.965 "data_offset": 2048, 00:10:46.965 "data_size": 63488 00:10:46.965 } 00:10:46.965 ] 00:10:46.965 } 00:10:46.965 } 00:10:46.965 }' 00:10:46.966 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:47.225 BaseBdev2 00:10:47.225 BaseBdev3 00:10:47.225 BaseBdev4' 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.225 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.225 [2024-12-05 19:01:04.724022] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:47.226 "name": "Existed_Raid", 00:10:47.226 "uuid": "23ed8fe5-5c6e-45df-a9de-cb82c9b30d5f", 00:10:47.226 "strip_size_kb": 0, 00:10:47.226 "state": "online", 00:10:47.226 "raid_level": "raid1", 00:10:47.226 "superblock": true, 00:10:47.226 "num_base_bdevs": 4, 00:10:47.226 "num_base_bdevs_discovered": 3, 00:10:47.226 "num_base_bdevs_operational": 3, 00:10:47.226 "base_bdevs_list": [ 00:10:47.226 { 00:10:47.226 "name": null, 00:10:47.226 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:47.226 "is_configured": false, 00:10:47.226 "data_offset": 0, 00:10:47.226 "data_size": 63488 00:10:47.226 }, 00:10:47.226 { 00:10:47.226 "name": "BaseBdev2", 00:10:47.226 "uuid": "9cf7245e-b1af-417d-98bd-539129a83e4b", 00:10:47.226 "is_configured": true, 00:10:47.226 "data_offset": 2048, 00:10:47.226 "data_size": 63488 00:10:47.226 }, 00:10:47.226 { 00:10:47.226 "name": "BaseBdev3", 00:10:47.226 "uuid": "efc36ed2-a2be-4d31-9cd9-2a77b4c31e8b", 00:10:47.226 "is_configured": true, 00:10:47.226 "data_offset": 2048, 00:10:47.226 "data_size": 63488 00:10:47.226 }, 00:10:47.226 { 00:10:47.226 "name": "BaseBdev4", 00:10:47.226 "uuid": "6a69c794-5070-420e-9852-bbd207e6c6d7", 00:10:47.226 "is_configured": true, 00:10:47.226 "data_offset": 2048, 00:10:47.226 "data_size": 63488 00:10:47.226 } 00:10:47.226 ] 00:10:47.226 }' 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:47.226 19:01:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.794 [2024-12-05 19:01:05.198353] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.794 [2024-12-05 19:01:05.269557] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.794 [2024-12-05 19:01:05.324701] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:47.794 [2024-12-05 19:01:05.324847] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:47.794 [2024-12-05 19:01:05.336300] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:47.794 [2024-12-05 19:01:05.336421] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:47.794 [2024-12-05 19:01:05.336463] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:47.794 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.072 BaseBdev2 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.072 [ 00:10:48.072 { 00:10:48.072 "name": "BaseBdev2", 00:10:48.072 "aliases": [ 00:10:48.072 "4a8eeda4-6eb6-4578-ab5d-b1d0312dd918" 00:10:48.072 ], 00:10:48.072 "product_name": "Malloc disk", 00:10:48.072 "block_size": 512, 00:10:48.072 "num_blocks": 65536, 00:10:48.072 "uuid": "4a8eeda4-6eb6-4578-ab5d-b1d0312dd918", 00:10:48.072 "assigned_rate_limits": { 00:10:48.072 "rw_ios_per_sec": 0, 00:10:48.072 "rw_mbytes_per_sec": 0, 00:10:48.072 "r_mbytes_per_sec": 0, 00:10:48.072 "w_mbytes_per_sec": 0 00:10:48.072 }, 00:10:48.072 "claimed": false, 00:10:48.072 "zoned": false, 00:10:48.072 "supported_io_types": { 00:10:48.072 "read": true, 00:10:48.072 "write": true, 00:10:48.072 "unmap": true, 00:10:48.072 "flush": true, 00:10:48.072 "reset": true, 00:10:48.072 "nvme_admin": false, 00:10:48.072 "nvme_io": false, 00:10:48.072 "nvme_io_md": false, 00:10:48.072 "write_zeroes": true, 00:10:48.072 "zcopy": true, 00:10:48.072 "get_zone_info": false, 00:10:48.072 "zone_management": false, 00:10:48.072 "zone_append": false, 00:10:48.072 "compare": false, 00:10:48.072 "compare_and_write": false, 00:10:48.072 "abort": true, 00:10:48.072 "seek_hole": false, 00:10:48.072 "seek_data": false, 00:10:48.072 "copy": true, 00:10:48.072 "nvme_iov_md": false 00:10:48.072 }, 00:10:48.072 "memory_domains": [ 00:10:48.072 { 00:10:48.072 "dma_device_id": "system", 00:10:48.072 "dma_device_type": 1 00:10:48.072 }, 00:10:48.072 { 00:10:48.072 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:48.072 "dma_device_type": 2 00:10:48.072 } 00:10:48.072 ], 00:10:48.072 "driver_specific": {} 00:10:48.072 } 00:10:48.072 ] 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:48.072 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.073 BaseBdev3 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.073 [ 00:10:48.073 { 00:10:48.073 "name": "BaseBdev3", 00:10:48.073 "aliases": [ 00:10:48.073 "ab829ffe-7c5f-4a0a-9785-c80b53816e12" 00:10:48.073 ], 00:10:48.073 "product_name": "Malloc disk", 00:10:48.073 "block_size": 512, 00:10:48.073 "num_blocks": 65536, 00:10:48.073 "uuid": "ab829ffe-7c5f-4a0a-9785-c80b53816e12", 00:10:48.073 "assigned_rate_limits": { 00:10:48.073 "rw_ios_per_sec": 0, 00:10:48.073 "rw_mbytes_per_sec": 0, 00:10:48.073 "r_mbytes_per_sec": 0, 00:10:48.073 "w_mbytes_per_sec": 0 00:10:48.073 }, 00:10:48.073 "claimed": false, 00:10:48.073 "zoned": false, 00:10:48.073 "supported_io_types": { 00:10:48.073 "read": true, 00:10:48.073 "write": true, 00:10:48.073 "unmap": true, 00:10:48.073 "flush": true, 00:10:48.073 "reset": true, 00:10:48.073 "nvme_admin": false, 00:10:48.073 "nvme_io": false, 00:10:48.073 "nvme_io_md": false, 00:10:48.073 "write_zeroes": true, 00:10:48.073 "zcopy": true, 00:10:48.073 "get_zone_info": false, 00:10:48.073 "zone_management": false, 00:10:48.073 "zone_append": false, 00:10:48.073 "compare": false, 00:10:48.073 "compare_and_write": false, 00:10:48.073 "abort": true, 00:10:48.073 "seek_hole": false, 00:10:48.073 "seek_data": false, 00:10:48.073 "copy": true, 00:10:48.073 "nvme_iov_md": false 00:10:48.073 }, 00:10:48.073 "memory_domains": [ 00:10:48.073 { 00:10:48.073 "dma_device_id": "system", 00:10:48.073 "dma_device_type": 1 00:10:48.073 }, 00:10:48.073 { 00:10:48.073 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:48.073 "dma_device_type": 2 00:10:48.073 } 00:10:48.073 ], 00:10:48.073 "driver_specific": {} 00:10:48.073 } 00:10:48.073 ] 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.073 BaseBdev4 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.073 [ 00:10:48.073 { 00:10:48.073 "name": "BaseBdev4", 00:10:48.073 "aliases": [ 00:10:48.073 "25e2d77b-f459-467f-8c17-8b95a2a59bb4" 00:10:48.073 ], 00:10:48.073 "product_name": "Malloc disk", 00:10:48.073 "block_size": 512, 00:10:48.073 "num_blocks": 65536, 00:10:48.073 "uuid": "25e2d77b-f459-467f-8c17-8b95a2a59bb4", 00:10:48.073 "assigned_rate_limits": { 00:10:48.073 "rw_ios_per_sec": 0, 00:10:48.073 "rw_mbytes_per_sec": 0, 00:10:48.073 "r_mbytes_per_sec": 0, 00:10:48.073 "w_mbytes_per_sec": 0 00:10:48.073 }, 00:10:48.073 "claimed": false, 00:10:48.073 "zoned": false, 00:10:48.073 "supported_io_types": { 00:10:48.073 "read": true, 00:10:48.073 "write": true, 00:10:48.073 "unmap": true, 00:10:48.073 "flush": true, 00:10:48.073 "reset": true, 00:10:48.073 "nvme_admin": false, 00:10:48.073 "nvme_io": false, 00:10:48.073 "nvme_io_md": false, 00:10:48.073 "write_zeroes": true, 00:10:48.073 "zcopy": true, 00:10:48.073 "get_zone_info": false, 00:10:48.073 "zone_management": false, 00:10:48.073 "zone_append": false, 00:10:48.073 "compare": false, 00:10:48.073 "compare_and_write": false, 00:10:48.073 "abort": true, 00:10:48.073 "seek_hole": false, 00:10:48.073 "seek_data": false, 00:10:48.073 "copy": true, 00:10:48.073 "nvme_iov_md": false 00:10:48.073 }, 00:10:48.073 "memory_domains": [ 00:10:48.073 { 00:10:48.073 "dma_device_id": "system", 00:10:48.073 "dma_device_type": 1 00:10:48.073 }, 00:10:48.073 { 00:10:48.073 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:48.073 "dma_device_type": 2 00:10:48.073 } 00:10:48.073 ], 00:10:48.073 "driver_specific": {} 00:10:48.073 } 00:10:48.073 ] 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.073 [2024-12-05 19:01:05.544472] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:48.073 [2024-12-05 19:01:05.544580] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:48.073 [2024-12-05 19:01:05.544619] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:48.073 [2024-12-05 19:01:05.546411] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:48.073 [2024-12-05 19:01:05.546503] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:48.073 "name": "Existed_Raid", 00:10:48.073 "uuid": "86740a63-a5b9-4e33-b4b1-9450ca635616", 00:10:48.073 "strip_size_kb": 0, 00:10:48.073 "state": "configuring", 00:10:48.073 "raid_level": "raid1", 00:10:48.073 "superblock": true, 00:10:48.073 "num_base_bdevs": 4, 00:10:48.073 "num_base_bdevs_discovered": 3, 00:10:48.073 "num_base_bdevs_operational": 4, 00:10:48.073 "base_bdevs_list": [ 00:10:48.073 { 00:10:48.073 "name": "BaseBdev1", 00:10:48.073 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:48.073 "is_configured": false, 00:10:48.073 "data_offset": 0, 00:10:48.073 "data_size": 0 00:10:48.073 }, 00:10:48.073 { 00:10:48.073 "name": "BaseBdev2", 00:10:48.073 "uuid": "4a8eeda4-6eb6-4578-ab5d-b1d0312dd918", 00:10:48.073 "is_configured": true, 00:10:48.073 "data_offset": 2048, 00:10:48.073 "data_size": 63488 00:10:48.073 }, 00:10:48.073 { 00:10:48.073 "name": "BaseBdev3", 00:10:48.073 "uuid": "ab829ffe-7c5f-4a0a-9785-c80b53816e12", 00:10:48.073 "is_configured": true, 00:10:48.073 "data_offset": 2048, 00:10:48.073 "data_size": 63488 00:10:48.073 }, 00:10:48.073 { 00:10:48.073 "name": "BaseBdev4", 00:10:48.073 "uuid": "25e2d77b-f459-467f-8c17-8b95a2a59bb4", 00:10:48.073 "is_configured": true, 00:10:48.073 "data_offset": 2048, 00:10:48.073 "data_size": 63488 00:10:48.073 } 00:10:48.073 ] 00:10:48.073 }' 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:48.073 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.642 [2024-12-05 19:01:05.983769] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:48.642 19:01:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.642 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.642 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:48.642 "name": "Existed_Raid", 00:10:48.642 "uuid": "86740a63-a5b9-4e33-b4b1-9450ca635616", 00:10:48.642 "strip_size_kb": 0, 00:10:48.642 "state": "configuring", 00:10:48.642 "raid_level": "raid1", 00:10:48.642 "superblock": true, 00:10:48.642 "num_base_bdevs": 4, 00:10:48.642 "num_base_bdevs_discovered": 2, 00:10:48.642 "num_base_bdevs_operational": 4, 00:10:48.642 "base_bdevs_list": [ 00:10:48.642 { 00:10:48.642 "name": "BaseBdev1", 00:10:48.642 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:48.642 "is_configured": false, 00:10:48.642 "data_offset": 0, 00:10:48.642 "data_size": 0 00:10:48.642 }, 00:10:48.642 { 00:10:48.642 "name": null, 00:10:48.642 "uuid": "4a8eeda4-6eb6-4578-ab5d-b1d0312dd918", 00:10:48.642 "is_configured": false, 00:10:48.642 "data_offset": 0, 00:10:48.642 "data_size": 63488 00:10:48.642 }, 00:10:48.642 { 00:10:48.642 "name": "BaseBdev3", 00:10:48.642 "uuid": "ab829ffe-7c5f-4a0a-9785-c80b53816e12", 00:10:48.642 "is_configured": true, 00:10:48.642 "data_offset": 2048, 00:10:48.642 "data_size": 63488 00:10:48.642 }, 00:10:48.642 { 00:10:48.642 "name": "BaseBdev4", 00:10:48.642 "uuid": "25e2d77b-f459-467f-8c17-8b95a2a59bb4", 00:10:48.642 "is_configured": true, 00:10:48.642 "data_offset": 2048, 00:10:48.642 "data_size": 63488 00:10:48.642 } 00:10:48.642 ] 00:10:48.642 }' 00:10:48.642 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:48.642 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.902 BaseBdev1 00:10:48.902 [2024-12-05 19:01:06.450111] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:48.902 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.162 [ 00:10:49.162 { 00:10:49.162 "name": "BaseBdev1", 00:10:49.162 "aliases": [ 00:10:49.162 "ee2f2c56-87a9-44fb-8e8d-08d6f7cc4361" 00:10:49.162 ], 00:10:49.162 "product_name": "Malloc disk", 00:10:49.162 "block_size": 512, 00:10:49.162 "num_blocks": 65536, 00:10:49.162 "uuid": "ee2f2c56-87a9-44fb-8e8d-08d6f7cc4361", 00:10:49.162 "assigned_rate_limits": { 00:10:49.162 "rw_ios_per_sec": 0, 00:10:49.162 "rw_mbytes_per_sec": 0, 00:10:49.162 "r_mbytes_per_sec": 0, 00:10:49.162 "w_mbytes_per_sec": 0 00:10:49.162 }, 00:10:49.162 "claimed": true, 00:10:49.162 "claim_type": "exclusive_write", 00:10:49.162 "zoned": false, 00:10:49.162 "supported_io_types": { 00:10:49.162 "read": true, 00:10:49.162 "write": true, 00:10:49.162 "unmap": true, 00:10:49.162 "flush": true, 00:10:49.162 "reset": true, 00:10:49.162 "nvme_admin": false, 00:10:49.162 "nvme_io": false, 00:10:49.162 "nvme_io_md": false, 00:10:49.162 "write_zeroes": true, 00:10:49.162 "zcopy": true, 00:10:49.162 "get_zone_info": false, 00:10:49.162 "zone_management": false, 00:10:49.162 "zone_append": false, 00:10:49.162 "compare": false, 00:10:49.162 "compare_and_write": false, 00:10:49.162 "abort": true, 00:10:49.162 "seek_hole": false, 00:10:49.162 "seek_data": false, 00:10:49.162 "copy": true, 00:10:49.162 "nvme_iov_md": false 00:10:49.162 }, 00:10:49.162 "memory_domains": [ 00:10:49.162 { 00:10:49.162 "dma_device_id": "system", 00:10:49.162 "dma_device_type": 1 00:10:49.162 }, 00:10:49.162 { 00:10:49.162 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:49.162 "dma_device_type": 2 00:10:49.162 } 00:10:49.162 ], 00:10:49.162 "driver_specific": {} 00:10:49.162 } 00:10:49.162 ] 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.162 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:49.162 "name": "Existed_Raid", 00:10:49.162 "uuid": "86740a63-a5b9-4e33-b4b1-9450ca635616", 00:10:49.162 "strip_size_kb": 0, 00:10:49.162 "state": "configuring", 00:10:49.162 "raid_level": "raid1", 00:10:49.162 "superblock": true, 00:10:49.162 "num_base_bdevs": 4, 00:10:49.162 "num_base_bdevs_discovered": 3, 00:10:49.162 "num_base_bdevs_operational": 4, 00:10:49.162 "base_bdevs_list": [ 00:10:49.162 { 00:10:49.162 "name": "BaseBdev1", 00:10:49.162 "uuid": "ee2f2c56-87a9-44fb-8e8d-08d6f7cc4361", 00:10:49.162 "is_configured": true, 00:10:49.162 "data_offset": 2048, 00:10:49.162 "data_size": 63488 00:10:49.162 }, 00:10:49.162 { 00:10:49.162 "name": null, 00:10:49.162 "uuid": "4a8eeda4-6eb6-4578-ab5d-b1d0312dd918", 00:10:49.162 "is_configured": false, 00:10:49.162 "data_offset": 0, 00:10:49.162 "data_size": 63488 00:10:49.163 }, 00:10:49.163 { 00:10:49.163 "name": "BaseBdev3", 00:10:49.163 "uuid": "ab829ffe-7c5f-4a0a-9785-c80b53816e12", 00:10:49.163 "is_configured": true, 00:10:49.163 "data_offset": 2048, 00:10:49.163 "data_size": 63488 00:10:49.163 }, 00:10:49.163 { 00:10:49.163 "name": "BaseBdev4", 00:10:49.163 "uuid": "25e2d77b-f459-467f-8c17-8b95a2a59bb4", 00:10:49.163 "is_configured": true, 00:10:49.163 "data_offset": 2048, 00:10:49.163 "data_size": 63488 00:10:49.163 } 00:10:49.163 ] 00:10:49.163 }' 00:10:49.163 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:49.163 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.423 [2024-12-05 19:01:06.933321] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:49.423 "name": "Existed_Raid", 00:10:49.423 "uuid": "86740a63-a5b9-4e33-b4b1-9450ca635616", 00:10:49.423 "strip_size_kb": 0, 00:10:49.423 "state": "configuring", 00:10:49.423 "raid_level": "raid1", 00:10:49.423 "superblock": true, 00:10:49.423 "num_base_bdevs": 4, 00:10:49.423 "num_base_bdevs_discovered": 2, 00:10:49.423 "num_base_bdevs_operational": 4, 00:10:49.423 "base_bdevs_list": [ 00:10:49.423 { 00:10:49.423 "name": "BaseBdev1", 00:10:49.423 "uuid": "ee2f2c56-87a9-44fb-8e8d-08d6f7cc4361", 00:10:49.423 "is_configured": true, 00:10:49.423 "data_offset": 2048, 00:10:49.423 "data_size": 63488 00:10:49.423 }, 00:10:49.423 { 00:10:49.423 "name": null, 00:10:49.423 "uuid": "4a8eeda4-6eb6-4578-ab5d-b1d0312dd918", 00:10:49.423 "is_configured": false, 00:10:49.423 "data_offset": 0, 00:10:49.423 "data_size": 63488 00:10:49.423 }, 00:10:49.423 { 00:10:49.423 "name": null, 00:10:49.423 "uuid": "ab829ffe-7c5f-4a0a-9785-c80b53816e12", 00:10:49.423 "is_configured": false, 00:10:49.423 "data_offset": 0, 00:10:49.423 "data_size": 63488 00:10:49.423 }, 00:10:49.423 { 00:10:49.423 "name": "BaseBdev4", 00:10:49.423 "uuid": "25e2d77b-f459-467f-8c17-8b95a2a59bb4", 00:10:49.423 "is_configured": true, 00:10:49.423 "data_offset": 2048, 00:10:49.423 "data_size": 63488 00:10:49.423 } 00:10:49.423 ] 00:10:49.423 }' 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:49.423 19:01:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.993 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.993 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.993 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.993 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:49.993 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.993 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:49.993 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:49.993 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.994 [2024-12-05 19:01:07.400571] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:49.994 "name": "Existed_Raid", 00:10:49.994 "uuid": "86740a63-a5b9-4e33-b4b1-9450ca635616", 00:10:49.994 "strip_size_kb": 0, 00:10:49.994 "state": "configuring", 00:10:49.994 "raid_level": "raid1", 00:10:49.994 "superblock": true, 00:10:49.994 "num_base_bdevs": 4, 00:10:49.994 "num_base_bdevs_discovered": 3, 00:10:49.994 "num_base_bdevs_operational": 4, 00:10:49.994 "base_bdevs_list": [ 00:10:49.994 { 00:10:49.994 "name": "BaseBdev1", 00:10:49.994 "uuid": "ee2f2c56-87a9-44fb-8e8d-08d6f7cc4361", 00:10:49.994 "is_configured": true, 00:10:49.994 "data_offset": 2048, 00:10:49.994 "data_size": 63488 00:10:49.994 }, 00:10:49.994 { 00:10:49.994 "name": null, 00:10:49.994 "uuid": "4a8eeda4-6eb6-4578-ab5d-b1d0312dd918", 00:10:49.994 "is_configured": false, 00:10:49.994 "data_offset": 0, 00:10:49.994 "data_size": 63488 00:10:49.994 }, 00:10:49.994 { 00:10:49.994 "name": "BaseBdev3", 00:10:49.994 "uuid": "ab829ffe-7c5f-4a0a-9785-c80b53816e12", 00:10:49.994 "is_configured": true, 00:10:49.994 "data_offset": 2048, 00:10:49.994 "data_size": 63488 00:10:49.994 }, 00:10:49.994 { 00:10:49.994 "name": "BaseBdev4", 00:10:49.994 "uuid": "25e2d77b-f459-467f-8c17-8b95a2a59bb4", 00:10:49.994 "is_configured": true, 00:10:49.994 "data_offset": 2048, 00:10:49.994 "data_size": 63488 00:10:49.994 } 00:10:49.994 ] 00:10:49.994 }' 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:49.994 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.566 [2024-12-05 19:01:07.875877] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:50.566 "name": "Existed_Raid", 00:10:50.566 "uuid": "86740a63-a5b9-4e33-b4b1-9450ca635616", 00:10:50.566 "strip_size_kb": 0, 00:10:50.566 "state": "configuring", 00:10:50.566 "raid_level": "raid1", 00:10:50.566 "superblock": true, 00:10:50.566 "num_base_bdevs": 4, 00:10:50.566 "num_base_bdevs_discovered": 2, 00:10:50.566 "num_base_bdevs_operational": 4, 00:10:50.566 "base_bdevs_list": [ 00:10:50.566 { 00:10:50.566 "name": null, 00:10:50.566 "uuid": "ee2f2c56-87a9-44fb-8e8d-08d6f7cc4361", 00:10:50.566 "is_configured": false, 00:10:50.566 "data_offset": 0, 00:10:50.566 "data_size": 63488 00:10:50.566 }, 00:10:50.566 { 00:10:50.566 "name": null, 00:10:50.566 "uuid": "4a8eeda4-6eb6-4578-ab5d-b1d0312dd918", 00:10:50.566 "is_configured": false, 00:10:50.566 "data_offset": 0, 00:10:50.566 "data_size": 63488 00:10:50.566 }, 00:10:50.566 { 00:10:50.566 "name": "BaseBdev3", 00:10:50.566 "uuid": "ab829ffe-7c5f-4a0a-9785-c80b53816e12", 00:10:50.566 "is_configured": true, 00:10:50.566 "data_offset": 2048, 00:10:50.566 "data_size": 63488 00:10:50.566 }, 00:10:50.566 { 00:10:50.566 "name": "BaseBdev4", 00:10:50.566 "uuid": "25e2d77b-f459-467f-8c17-8b95a2a59bb4", 00:10:50.566 "is_configured": true, 00:10:50.566 "data_offset": 2048, 00:10:50.566 "data_size": 63488 00:10:50.566 } 00:10:50.566 ] 00:10:50.566 }' 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:50.566 19:01:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.826 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.827 [2024-12-05 19:01:08.309503] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:50.827 "name": "Existed_Raid", 00:10:50.827 "uuid": "86740a63-a5b9-4e33-b4b1-9450ca635616", 00:10:50.827 "strip_size_kb": 0, 00:10:50.827 "state": "configuring", 00:10:50.827 "raid_level": "raid1", 00:10:50.827 "superblock": true, 00:10:50.827 "num_base_bdevs": 4, 00:10:50.827 "num_base_bdevs_discovered": 3, 00:10:50.827 "num_base_bdevs_operational": 4, 00:10:50.827 "base_bdevs_list": [ 00:10:50.827 { 00:10:50.827 "name": null, 00:10:50.827 "uuid": "ee2f2c56-87a9-44fb-8e8d-08d6f7cc4361", 00:10:50.827 "is_configured": false, 00:10:50.827 "data_offset": 0, 00:10:50.827 "data_size": 63488 00:10:50.827 }, 00:10:50.827 { 00:10:50.827 "name": "BaseBdev2", 00:10:50.827 "uuid": "4a8eeda4-6eb6-4578-ab5d-b1d0312dd918", 00:10:50.827 "is_configured": true, 00:10:50.827 "data_offset": 2048, 00:10:50.827 "data_size": 63488 00:10:50.827 }, 00:10:50.827 { 00:10:50.827 "name": "BaseBdev3", 00:10:50.827 "uuid": "ab829ffe-7c5f-4a0a-9785-c80b53816e12", 00:10:50.827 "is_configured": true, 00:10:50.827 "data_offset": 2048, 00:10:50.827 "data_size": 63488 00:10:50.827 }, 00:10:50.827 { 00:10:50.827 "name": "BaseBdev4", 00:10:50.827 "uuid": "25e2d77b-f459-467f-8c17-8b95a2a59bb4", 00:10:50.827 "is_configured": true, 00:10:50.827 "data_offset": 2048, 00:10:50.827 "data_size": 63488 00:10:50.827 } 00:10:50.827 ] 00:10:50.827 }' 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:50.827 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.397 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:51.397 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:51.397 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.397 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.397 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.397 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:51.397 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:51.397 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:51.397 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.397 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.397 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u ee2f2c56-87a9-44fb-8e8d-08d6f7cc4361 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.398 [2024-12-05 19:01:08.795578] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:51.398 [2024-12-05 19:01:08.795762] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:51.398 [2024-12-05 19:01:08.795782] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:51.398 [2024-12-05 19:01:08.796046] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:51.398 [2024-12-05 19:01:08.796185] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:51.398 [2024-12-05 19:01:08.796195] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:51.398 [2024-12-05 19:01:08.796294] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:51.398 NewBaseBdev 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.398 [ 00:10:51.398 { 00:10:51.398 "name": "NewBaseBdev", 00:10:51.398 "aliases": [ 00:10:51.398 "ee2f2c56-87a9-44fb-8e8d-08d6f7cc4361" 00:10:51.398 ], 00:10:51.398 "product_name": "Malloc disk", 00:10:51.398 "block_size": 512, 00:10:51.398 "num_blocks": 65536, 00:10:51.398 "uuid": "ee2f2c56-87a9-44fb-8e8d-08d6f7cc4361", 00:10:51.398 "assigned_rate_limits": { 00:10:51.398 "rw_ios_per_sec": 0, 00:10:51.398 "rw_mbytes_per_sec": 0, 00:10:51.398 "r_mbytes_per_sec": 0, 00:10:51.398 "w_mbytes_per_sec": 0 00:10:51.398 }, 00:10:51.398 "claimed": true, 00:10:51.398 "claim_type": "exclusive_write", 00:10:51.398 "zoned": false, 00:10:51.398 "supported_io_types": { 00:10:51.398 "read": true, 00:10:51.398 "write": true, 00:10:51.398 "unmap": true, 00:10:51.398 "flush": true, 00:10:51.398 "reset": true, 00:10:51.398 "nvme_admin": false, 00:10:51.398 "nvme_io": false, 00:10:51.398 "nvme_io_md": false, 00:10:51.398 "write_zeroes": true, 00:10:51.398 "zcopy": true, 00:10:51.398 "get_zone_info": false, 00:10:51.398 "zone_management": false, 00:10:51.398 "zone_append": false, 00:10:51.398 "compare": false, 00:10:51.398 "compare_and_write": false, 00:10:51.398 "abort": true, 00:10:51.398 "seek_hole": false, 00:10:51.398 "seek_data": false, 00:10:51.398 "copy": true, 00:10:51.398 "nvme_iov_md": false 00:10:51.398 }, 00:10:51.398 "memory_domains": [ 00:10:51.398 { 00:10:51.398 "dma_device_id": "system", 00:10:51.398 "dma_device_type": 1 00:10:51.398 }, 00:10:51.398 { 00:10:51.398 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:51.398 "dma_device_type": 2 00:10:51.398 } 00:10:51.398 ], 00:10:51.398 "driver_specific": {} 00:10:51.398 } 00:10:51.398 ] 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:51.398 "name": "Existed_Raid", 00:10:51.398 "uuid": "86740a63-a5b9-4e33-b4b1-9450ca635616", 00:10:51.398 "strip_size_kb": 0, 00:10:51.398 "state": "online", 00:10:51.398 "raid_level": "raid1", 00:10:51.398 "superblock": true, 00:10:51.398 "num_base_bdevs": 4, 00:10:51.398 "num_base_bdevs_discovered": 4, 00:10:51.398 "num_base_bdevs_operational": 4, 00:10:51.398 "base_bdevs_list": [ 00:10:51.398 { 00:10:51.398 "name": "NewBaseBdev", 00:10:51.398 "uuid": "ee2f2c56-87a9-44fb-8e8d-08d6f7cc4361", 00:10:51.398 "is_configured": true, 00:10:51.398 "data_offset": 2048, 00:10:51.398 "data_size": 63488 00:10:51.398 }, 00:10:51.398 { 00:10:51.398 "name": "BaseBdev2", 00:10:51.398 "uuid": "4a8eeda4-6eb6-4578-ab5d-b1d0312dd918", 00:10:51.398 "is_configured": true, 00:10:51.398 "data_offset": 2048, 00:10:51.398 "data_size": 63488 00:10:51.398 }, 00:10:51.398 { 00:10:51.398 "name": "BaseBdev3", 00:10:51.398 "uuid": "ab829ffe-7c5f-4a0a-9785-c80b53816e12", 00:10:51.398 "is_configured": true, 00:10:51.398 "data_offset": 2048, 00:10:51.398 "data_size": 63488 00:10:51.398 }, 00:10:51.398 { 00:10:51.398 "name": "BaseBdev4", 00:10:51.398 "uuid": "25e2d77b-f459-467f-8c17-8b95a2a59bb4", 00:10:51.398 "is_configured": true, 00:10:51.398 "data_offset": 2048, 00:10:51.398 "data_size": 63488 00:10:51.398 } 00:10:51.398 ] 00:10:51.398 }' 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:51.398 19:01:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.968 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:51.968 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:51.969 [2024-12-05 19:01:09.243118] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:51.969 "name": "Existed_Raid", 00:10:51.969 "aliases": [ 00:10:51.969 "86740a63-a5b9-4e33-b4b1-9450ca635616" 00:10:51.969 ], 00:10:51.969 "product_name": "Raid Volume", 00:10:51.969 "block_size": 512, 00:10:51.969 "num_blocks": 63488, 00:10:51.969 "uuid": "86740a63-a5b9-4e33-b4b1-9450ca635616", 00:10:51.969 "assigned_rate_limits": { 00:10:51.969 "rw_ios_per_sec": 0, 00:10:51.969 "rw_mbytes_per_sec": 0, 00:10:51.969 "r_mbytes_per_sec": 0, 00:10:51.969 "w_mbytes_per_sec": 0 00:10:51.969 }, 00:10:51.969 "claimed": false, 00:10:51.969 "zoned": false, 00:10:51.969 "supported_io_types": { 00:10:51.969 "read": true, 00:10:51.969 "write": true, 00:10:51.969 "unmap": false, 00:10:51.969 "flush": false, 00:10:51.969 "reset": true, 00:10:51.969 "nvme_admin": false, 00:10:51.969 "nvme_io": false, 00:10:51.969 "nvme_io_md": false, 00:10:51.969 "write_zeroes": true, 00:10:51.969 "zcopy": false, 00:10:51.969 "get_zone_info": false, 00:10:51.969 "zone_management": false, 00:10:51.969 "zone_append": false, 00:10:51.969 "compare": false, 00:10:51.969 "compare_and_write": false, 00:10:51.969 "abort": false, 00:10:51.969 "seek_hole": false, 00:10:51.969 "seek_data": false, 00:10:51.969 "copy": false, 00:10:51.969 "nvme_iov_md": false 00:10:51.969 }, 00:10:51.969 "memory_domains": [ 00:10:51.969 { 00:10:51.969 "dma_device_id": "system", 00:10:51.969 "dma_device_type": 1 00:10:51.969 }, 00:10:51.969 { 00:10:51.969 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:51.969 "dma_device_type": 2 00:10:51.969 }, 00:10:51.969 { 00:10:51.969 "dma_device_id": "system", 00:10:51.969 "dma_device_type": 1 00:10:51.969 }, 00:10:51.969 { 00:10:51.969 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:51.969 "dma_device_type": 2 00:10:51.969 }, 00:10:51.969 { 00:10:51.969 "dma_device_id": "system", 00:10:51.969 "dma_device_type": 1 00:10:51.969 }, 00:10:51.969 { 00:10:51.969 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:51.969 "dma_device_type": 2 00:10:51.969 }, 00:10:51.969 { 00:10:51.969 "dma_device_id": "system", 00:10:51.969 "dma_device_type": 1 00:10:51.969 }, 00:10:51.969 { 00:10:51.969 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:51.969 "dma_device_type": 2 00:10:51.969 } 00:10:51.969 ], 00:10:51.969 "driver_specific": { 00:10:51.969 "raid": { 00:10:51.969 "uuid": "86740a63-a5b9-4e33-b4b1-9450ca635616", 00:10:51.969 "strip_size_kb": 0, 00:10:51.969 "state": "online", 00:10:51.969 "raid_level": "raid1", 00:10:51.969 "superblock": true, 00:10:51.969 "num_base_bdevs": 4, 00:10:51.969 "num_base_bdevs_discovered": 4, 00:10:51.969 "num_base_bdevs_operational": 4, 00:10:51.969 "base_bdevs_list": [ 00:10:51.969 { 00:10:51.969 "name": "NewBaseBdev", 00:10:51.969 "uuid": "ee2f2c56-87a9-44fb-8e8d-08d6f7cc4361", 00:10:51.969 "is_configured": true, 00:10:51.969 "data_offset": 2048, 00:10:51.969 "data_size": 63488 00:10:51.969 }, 00:10:51.969 { 00:10:51.969 "name": "BaseBdev2", 00:10:51.969 "uuid": "4a8eeda4-6eb6-4578-ab5d-b1d0312dd918", 00:10:51.969 "is_configured": true, 00:10:51.969 "data_offset": 2048, 00:10:51.969 "data_size": 63488 00:10:51.969 }, 00:10:51.969 { 00:10:51.969 "name": "BaseBdev3", 00:10:51.969 "uuid": "ab829ffe-7c5f-4a0a-9785-c80b53816e12", 00:10:51.969 "is_configured": true, 00:10:51.969 "data_offset": 2048, 00:10:51.969 "data_size": 63488 00:10:51.969 }, 00:10:51.969 { 00:10:51.969 "name": "BaseBdev4", 00:10:51.969 "uuid": "25e2d77b-f459-467f-8c17-8b95a2a59bb4", 00:10:51.969 "is_configured": true, 00:10:51.969 "data_offset": 2048, 00:10:51.969 "data_size": 63488 00:10:51.969 } 00:10:51.969 ] 00:10:51.969 } 00:10:51.969 } 00:10:51.969 }' 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:51.969 BaseBdev2 00:10:51.969 BaseBdev3 00:10:51.969 BaseBdev4' 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:51.969 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.230 [2024-12-05 19:01:09.574254] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:52.230 [2024-12-05 19:01:09.574282] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:52.230 [2024-12-05 19:01:09.574358] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:52.230 [2024-12-05 19:01:09.574626] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:52.230 [2024-12-05 19:01:09.574647] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 83521 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 83521 ']' 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 83521 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 83521 00:10:52.230 killing process with pid 83521 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 83521' 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 83521 00:10:52.230 [2024-12-05 19:01:09.611108] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:52.230 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 83521 00:10:52.230 [2024-12-05 19:01:09.650316] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:52.490 19:01:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:52.490 00:10:52.490 real 0m9.124s 00:10:52.490 user 0m15.613s 00:10:52.490 sys 0m1.851s 00:10:52.490 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:52.490 19:01:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:52.490 ************************************ 00:10:52.490 END TEST raid_state_function_test_sb 00:10:52.490 ************************************ 00:10:52.490 19:01:09 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 4 00:10:52.490 19:01:09 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:10:52.490 19:01:09 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:52.490 19:01:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:52.490 ************************************ 00:10:52.490 START TEST raid_superblock_test 00:10:52.490 ************************************ 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 4 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=84116 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 84116 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 84116 ']' 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:52.490 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:52.490 19:01:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:52.491 19:01:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:52.491 19:01:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.491 [2024-12-05 19:01:10.022835] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:10:52.491 [2024-12-05 19:01:10.022990] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid84116 ] 00:10:52.751 [2024-12-05 19:01:10.177459] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:52.751 [2024-12-05 19:01:10.202070] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:52.751 [2024-12-05 19:01:10.245004] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:52.751 [2024-12-05 19:01:10.245075] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.321 malloc1 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.321 [2024-12-05 19:01:10.857096] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:53.321 [2024-12-05 19:01:10.857195] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:53.321 [2024-12-05 19:01:10.857218] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:53.321 [2024-12-05 19:01:10.857238] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:53.321 [2024-12-05 19:01:10.859293] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:53.321 [2024-12-05 19:01:10.859331] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:53.321 pt1 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.321 malloc2 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.321 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.585 [2024-12-05 19:01:10.885790] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:53.585 [2024-12-05 19:01:10.885865] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:53.585 [2024-12-05 19:01:10.885883] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:53.585 [2024-12-05 19:01:10.885892] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:53.585 [2024-12-05 19:01:10.887955] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:53.585 [2024-12-05 19:01:10.887994] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:53.585 pt2 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.585 malloc3 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.585 [2024-12-05 19:01:10.914323] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:53.585 [2024-12-05 19:01:10.914395] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:53.585 [2024-12-05 19:01:10.914412] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:53.585 [2024-12-05 19:01:10.914422] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:53.585 [2024-12-05 19:01:10.916492] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:53.585 [2024-12-05 19:01:10.916533] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:53.585 pt3 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.585 malloc4 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.585 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.585 [2024-12-05 19:01:10.960325] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:53.585 [2024-12-05 19:01:10.960404] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:53.585 [2024-12-05 19:01:10.960430] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:53.586 [2024-12-05 19:01:10.960450] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:53.586 [2024-12-05 19:01:10.963717] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:53.586 [2024-12-05 19:01:10.963770] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:53.586 pt4 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.586 [2024-12-05 19:01:10.972395] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:53.586 [2024-12-05 19:01:10.974495] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:53.586 [2024-12-05 19:01:10.974576] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:53.586 [2024-12-05 19:01:10.974653] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:53.586 [2024-12-05 19:01:10.974848] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:53.586 [2024-12-05 19:01:10.974872] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:53.586 [2024-12-05 19:01:10.975147] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:53.586 [2024-12-05 19:01:10.975329] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:53.586 [2024-12-05 19:01:10.975348] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:53.586 [2024-12-05 19:01:10.975492] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.586 19:01:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.586 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:53.586 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:53.586 "name": "raid_bdev1", 00:10:53.586 "uuid": "665b16ec-9474-4950-a758-26cd1316d578", 00:10:53.586 "strip_size_kb": 0, 00:10:53.586 "state": "online", 00:10:53.586 "raid_level": "raid1", 00:10:53.586 "superblock": true, 00:10:53.586 "num_base_bdevs": 4, 00:10:53.586 "num_base_bdevs_discovered": 4, 00:10:53.586 "num_base_bdevs_operational": 4, 00:10:53.586 "base_bdevs_list": [ 00:10:53.586 { 00:10:53.586 "name": "pt1", 00:10:53.586 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:53.586 "is_configured": true, 00:10:53.586 "data_offset": 2048, 00:10:53.586 "data_size": 63488 00:10:53.586 }, 00:10:53.586 { 00:10:53.586 "name": "pt2", 00:10:53.586 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:53.586 "is_configured": true, 00:10:53.586 "data_offset": 2048, 00:10:53.586 "data_size": 63488 00:10:53.586 }, 00:10:53.586 { 00:10:53.586 "name": "pt3", 00:10:53.586 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:53.586 "is_configured": true, 00:10:53.586 "data_offset": 2048, 00:10:53.586 "data_size": 63488 00:10:53.586 }, 00:10:53.586 { 00:10:53.586 "name": "pt4", 00:10:53.586 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:53.586 "is_configured": true, 00:10:53.586 "data_offset": 2048, 00:10:53.586 "data_size": 63488 00:10:53.586 } 00:10:53.586 ] 00:10:53.586 }' 00:10:53.586 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:53.586 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.846 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:53.846 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:53.846 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:53.846 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:53.846 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:53.846 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:53.846 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:53.846 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:53.846 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:53.846 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.106 [2024-12-05 19:01:11.407959] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:54.106 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.106 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:54.106 "name": "raid_bdev1", 00:10:54.106 "aliases": [ 00:10:54.106 "665b16ec-9474-4950-a758-26cd1316d578" 00:10:54.106 ], 00:10:54.106 "product_name": "Raid Volume", 00:10:54.106 "block_size": 512, 00:10:54.106 "num_blocks": 63488, 00:10:54.106 "uuid": "665b16ec-9474-4950-a758-26cd1316d578", 00:10:54.106 "assigned_rate_limits": { 00:10:54.106 "rw_ios_per_sec": 0, 00:10:54.106 "rw_mbytes_per_sec": 0, 00:10:54.106 "r_mbytes_per_sec": 0, 00:10:54.106 "w_mbytes_per_sec": 0 00:10:54.106 }, 00:10:54.106 "claimed": false, 00:10:54.106 "zoned": false, 00:10:54.106 "supported_io_types": { 00:10:54.106 "read": true, 00:10:54.106 "write": true, 00:10:54.106 "unmap": false, 00:10:54.106 "flush": false, 00:10:54.106 "reset": true, 00:10:54.106 "nvme_admin": false, 00:10:54.106 "nvme_io": false, 00:10:54.106 "nvme_io_md": false, 00:10:54.106 "write_zeroes": true, 00:10:54.106 "zcopy": false, 00:10:54.106 "get_zone_info": false, 00:10:54.106 "zone_management": false, 00:10:54.106 "zone_append": false, 00:10:54.106 "compare": false, 00:10:54.106 "compare_and_write": false, 00:10:54.106 "abort": false, 00:10:54.106 "seek_hole": false, 00:10:54.106 "seek_data": false, 00:10:54.106 "copy": false, 00:10:54.107 "nvme_iov_md": false 00:10:54.107 }, 00:10:54.107 "memory_domains": [ 00:10:54.107 { 00:10:54.107 "dma_device_id": "system", 00:10:54.107 "dma_device_type": 1 00:10:54.107 }, 00:10:54.107 { 00:10:54.107 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:54.107 "dma_device_type": 2 00:10:54.107 }, 00:10:54.107 { 00:10:54.107 "dma_device_id": "system", 00:10:54.107 "dma_device_type": 1 00:10:54.107 }, 00:10:54.107 { 00:10:54.107 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:54.107 "dma_device_type": 2 00:10:54.107 }, 00:10:54.107 { 00:10:54.107 "dma_device_id": "system", 00:10:54.107 "dma_device_type": 1 00:10:54.107 }, 00:10:54.107 { 00:10:54.107 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:54.107 "dma_device_type": 2 00:10:54.107 }, 00:10:54.107 { 00:10:54.107 "dma_device_id": "system", 00:10:54.107 "dma_device_type": 1 00:10:54.107 }, 00:10:54.107 { 00:10:54.107 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:54.107 "dma_device_type": 2 00:10:54.107 } 00:10:54.107 ], 00:10:54.107 "driver_specific": { 00:10:54.107 "raid": { 00:10:54.107 "uuid": "665b16ec-9474-4950-a758-26cd1316d578", 00:10:54.107 "strip_size_kb": 0, 00:10:54.107 "state": "online", 00:10:54.107 "raid_level": "raid1", 00:10:54.107 "superblock": true, 00:10:54.107 "num_base_bdevs": 4, 00:10:54.107 "num_base_bdevs_discovered": 4, 00:10:54.107 "num_base_bdevs_operational": 4, 00:10:54.107 "base_bdevs_list": [ 00:10:54.107 { 00:10:54.107 "name": "pt1", 00:10:54.107 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:54.107 "is_configured": true, 00:10:54.107 "data_offset": 2048, 00:10:54.107 "data_size": 63488 00:10:54.107 }, 00:10:54.107 { 00:10:54.107 "name": "pt2", 00:10:54.107 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:54.107 "is_configured": true, 00:10:54.107 "data_offset": 2048, 00:10:54.107 "data_size": 63488 00:10:54.107 }, 00:10:54.107 { 00:10:54.107 "name": "pt3", 00:10:54.107 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:54.107 "is_configured": true, 00:10:54.107 "data_offset": 2048, 00:10:54.107 "data_size": 63488 00:10:54.107 }, 00:10:54.107 { 00:10:54.107 "name": "pt4", 00:10:54.107 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:54.107 "is_configured": true, 00:10:54.107 "data_offset": 2048, 00:10:54.107 "data_size": 63488 00:10:54.107 } 00:10:54.107 ] 00:10:54.107 } 00:10:54.107 } 00:10:54.107 }' 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:54.107 pt2 00:10:54.107 pt3 00:10:54.107 pt4' 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:54.107 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:54.370 [2024-12-05 19:01:11.707391] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=665b16ec-9474-4950-a758-26cd1316d578 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 665b16ec-9474-4950-a758-26cd1316d578 ']' 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.370 [2024-12-05 19:01:11.739064] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:54.370 [2024-12-05 19:01:11.739093] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:54.370 [2024-12-05 19:01:11.739165] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:54.370 [2024-12-05 19:01:11.739250] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:54.370 [2024-12-05 19:01:11.739260] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.370 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.371 [2024-12-05 19:01:11.902802] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:54.371 [2024-12-05 19:01:11.904622] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:54.371 [2024-12-05 19:01:11.904686] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:54.371 [2024-12-05 19:01:11.904716] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:10:54.371 [2024-12-05 19:01:11.904760] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:54.371 [2024-12-05 19:01:11.904799] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:54.371 [2024-12-05 19:01:11.904817] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:54.371 [2024-12-05 19:01:11.904833] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:10:54.371 [2024-12-05 19:01:11.904846] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:54.371 [2024-12-05 19:01:11.904863] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:10:54.371 request: 00:10:54.371 { 00:10:54.371 "name": "raid_bdev1", 00:10:54.371 "raid_level": "raid1", 00:10:54.371 "base_bdevs": [ 00:10:54.371 "malloc1", 00:10:54.371 "malloc2", 00:10:54.371 "malloc3", 00:10:54.371 "malloc4" 00:10:54.371 ], 00:10:54.371 "superblock": false, 00:10:54.371 "method": "bdev_raid_create", 00:10:54.371 "req_id": 1 00:10:54.371 } 00:10:54.371 Got JSON-RPC error response 00:10:54.371 response: 00:10:54.371 { 00:10:54.371 "code": -17, 00:10:54.371 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:54.371 } 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.371 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.631 [2024-12-05 19:01:11.954708] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:54.631 [2024-12-05 19:01:11.954768] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:54.631 [2024-12-05 19:01:11.954788] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:54.631 [2024-12-05 19:01:11.954796] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:54.631 [2024-12-05 19:01:11.956936] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:54.631 [2024-12-05 19:01:11.956974] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:54.631 [2024-12-05 19:01:11.957035] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:54.631 [2024-12-05 19:01:11.957071] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:54.631 pt1 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:54.631 19:01:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.631 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:54.631 "name": "raid_bdev1", 00:10:54.631 "uuid": "665b16ec-9474-4950-a758-26cd1316d578", 00:10:54.631 "strip_size_kb": 0, 00:10:54.631 "state": "configuring", 00:10:54.632 "raid_level": "raid1", 00:10:54.632 "superblock": true, 00:10:54.632 "num_base_bdevs": 4, 00:10:54.632 "num_base_bdevs_discovered": 1, 00:10:54.632 "num_base_bdevs_operational": 4, 00:10:54.632 "base_bdevs_list": [ 00:10:54.632 { 00:10:54.632 "name": "pt1", 00:10:54.632 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:54.632 "is_configured": true, 00:10:54.632 "data_offset": 2048, 00:10:54.632 "data_size": 63488 00:10:54.632 }, 00:10:54.632 { 00:10:54.632 "name": null, 00:10:54.632 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:54.632 "is_configured": false, 00:10:54.632 "data_offset": 2048, 00:10:54.632 "data_size": 63488 00:10:54.632 }, 00:10:54.632 { 00:10:54.632 "name": null, 00:10:54.632 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:54.632 "is_configured": false, 00:10:54.632 "data_offset": 2048, 00:10:54.632 "data_size": 63488 00:10:54.632 }, 00:10:54.632 { 00:10:54.632 "name": null, 00:10:54.632 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:54.632 "is_configured": false, 00:10:54.632 "data_offset": 2048, 00:10:54.632 "data_size": 63488 00:10:54.632 } 00:10:54.632 ] 00:10:54.632 }' 00:10:54.632 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:54.632 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.892 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:54.892 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:54.892 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.892 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.892 [2024-12-05 19:01:12.421870] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:54.892 [2024-12-05 19:01:12.421922] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:54.892 [2024-12-05 19:01:12.421939] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:10:54.893 [2024-12-05 19:01:12.421956] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:54.893 [2024-12-05 19:01:12.422296] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:54.893 [2024-12-05 19:01:12.422320] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:54.893 [2024-12-05 19:01:12.422381] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:54.893 [2024-12-05 19:01:12.422404] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:54.893 pt2 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.893 [2024-12-05 19:01:12.433877] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:54.893 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.153 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.153 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:55.153 "name": "raid_bdev1", 00:10:55.153 "uuid": "665b16ec-9474-4950-a758-26cd1316d578", 00:10:55.153 "strip_size_kb": 0, 00:10:55.153 "state": "configuring", 00:10:55.153 "raid_level": "raid1", 00:10:55.153 "superblock": true, 00:10:55.153 "num_base_bdevs": 4, 00:10:55.153 "num_base_bdevs_discovered": 1, 00:10:55.153 "num_base_bdevs_operational": 4, 00:10:55.153 "base_bdevs_list": [ 00:10:55.153 { 00:10:55.153 "name": "pt1", 00:10:55.153 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:55.153 "is_configured": true, 00:10:55.153 "data_offset": 2048, 00:10:55.153 "data_size": 63488 00:10:55.153 }, 00:10:55.153 { 00:10:55.153 "name": null, 00:10:55.153 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:55.153 "is_configured": false, 00:10:55.153 "data_offset": 0, 00:10:55.153 "data_size": 63488 00:10:55.153 }, 00:10:55.153 { 00:10:55.153 "name": null, 00:10:55.153 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:55.153 "is_configured": false, 00:10:55.153 "data_offset": 2048, 00:10:55.153 "data_size": 63488 00:10:55.153 }, 00:10:55.153 { 00:10:55.153 "name": null, 00:10:55.153 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:55.153 "is_configured": false, 00:10:55.153 "data_offset": 2048, 00:10:55.153 "data_size": 63488 00:10:55.153 } 00:10:55.153 ] 00:10:55.153 }' 00:10:55.153 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:55.153 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.413 [2024-12-05 19:01:12.825224] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:55.413 [2024-12-05 19:01:12.825290] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:55.413 [2024-12-05 19:01:12.825306] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:10:55.413 [2024-12-05 19:01:12.825316] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:55.413 [2024-12-05 19:01:12.825696] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:55.413 [2024-12-05 19:01:12.825729] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:55.413 [2024-12-05 19:01:12.825794] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:55.413 [2024-12-05 19:01:12.825817] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:55.413 pt2 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.413 [2024-12-05 19:01:12.837167] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:55.413 [2024-12-05 19:01:12.837232] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:55.413 [2024-12-05 19:01:12.837246] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:55.413 [2024-12-05 19:01:12.837255] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:55.413 [2024-12-05 19:01:12.837581] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:55.413 [2024-12-05 19:01:12.837607] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:55.413 [2024-12-05 19:01:12.837657] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:55.413 [2024-12-05 19:01:12.837701] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:55.413 pt3 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.413 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.413 [2024-12-05 19:01:12.849148] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:55.413 [2024-12-05 19:01:12.849197] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:55.413 [2024-12-05 19:01:12.849210] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:10:55.413 [2024-12-05 19:01:12.849219] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:55.413 [2024-12-05 19:01:12.849499] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:55.413 [2024-12-05 19:01:12.849526] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:55.413 [2024-12-05 19:01:12.849576] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:55.413 [2024-12-05 19:01:12.849595] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:55.413 [2024-12-05 19:01:12.849718] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:55.414 [2024-12-05 19:01:12.849741] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:55.414 [2024-12-05 19:01:12.849962] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:10:55.414 [2024-12-05 19:01:12.850095] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:55.414 [2024-12-05 19:01:12.850108] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:10:55.414 [2024-12-05 19:01:12.850207] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:55.414 pt4 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:55.414 "name": "raid_bdev1", 00:10:55.414 "uuid": "665b16ec-9474-4950-a758-26cd1316d578", 00:10:55.414 "strip_size_kb": 0, 00:10:55.414 "state": "online", 00:10:55.414 "raid_level": "raid1", 00:10:55.414 "superblock": true, 00:10:55.414 "num_base_bdevs": 4, 00:10:55.414 "num_base_bdevs_discovered": 4, 00:10:55.414 "num_base_bdevs_operational": 4, 00:10:55.414 "base_bdevs_list": [ 00:10:55.414 { 00:10:55.414 "name": "pt1", 00:10:55.414 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:55.414 "is_configured": true, 00:10:55.414 "data_offset": 2048, 00:10:55.414 "data_size": 63488 00:10:55.414 }, 00:10:55.414 { 00:10:55.414 "name": "pt2", 00:10:55.414 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:55.414 "is_configured": true, 00:10:55.414 "data_offset": 2048, 00:10:55.414 "data_size": 63488 00:10:55.414 }, 00:10:55.414 { 00:10:55.414 "name": "pt3", 00:10:55.414 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:55.414 "is_configured": true, 00:10:55.414 "data_offset": 2048, 00:10:55.414 "data_size": 63488 00:10:55.414 }, 00:10:55.414 { 00:10:55.414 "name": "pt4", 00:10:55.414 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:55.414 "is_configured": true, 00:10:55.414 "data_offset": 2048, 00:10:55.414 "data_size": 63488 00:10:55.414 } 00:10:55.414 ] 00:10:55.414 }' 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:55.414 19:01:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.984 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:55.984 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:55.984 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:55.984 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:55.984 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:55.984 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:55.984 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:55.984 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:55.984 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.984 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.984 [2024-12-05 19:01:13.296706] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:55.984 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.984 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:55.984 "name": "raid_bdev1", 00:10:55.984 "aliases": [ 00:10:55.984 "665b16ec-9474-4950-a758-26cd1316d578" 00:10:55.984 ], 00:10:55.984 "product_name": "Raid Volume", 00:10:55.984 "block_size": 512, 00:10:55.984 "num_blocks": 63488, 00:10:55.984 "uuid": "665b16ec-9474-4950-a758-26cd1316d578", 00:10:55.984 "assigned_rate_limits": { 00:10:55.984 "rw_ios_per_sec": 0, 00:10:55.984 "rw_mbytes_per_sec": 0, 00:10:55.984 "r_mbytes_per_sec": 0, 00:10:55.984 "w_mbytes_per_sec": 0 00:10:55.984 }, 00:10:55.984 "claimed": false, 00:10:55.984 "zoned": false, 00:10:55.984 "supported_io_types": { 00:10:55.984 "read": true, 00:10:55.984 "write": true, 00:10:55.984 "unmap": false, 00:10:55.984 "flush": false, 00:10:55.984 "reset": true, 00:10:55.984 "nvme_admin": false, 00:10:55.984 "nvme_io": false, 00:10:55.984 "nvme_io_md": false, 00:10:55.984 "write_zeroes": true, 00:10:55.984 "zcopy": false, 00:10:55.984 "get_zone_info": false, 00:10:55.984 "zone_management": false, 00:10:55.984 "zone_append": false, 00:10:55.984 "compare": false, 00:10:55.984 "compare_and_write": false, 00:10:55.984 "abort": false, 00:10:55.984 "seek_hole": false, 00:10:55.984 "seek_data": false, 00:10:55.984 "copy": false, 00:10:55.984 "nvme_iov_md": false 00:10:55.984 }, 00:10:55.984 "memory_domains": [ 00:10:55.984 { 00:10:55.984 "dma_device_id": "system", 00:10:55.984 "dma_device_type": 1 00:10:55.984 }, 00:10:55.984 { 00:10:55.984 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:55.985 "dma_device_type": 2 00:10:55.985 }, 00:10:55.985 { 00:10:55.985 "dma_device_id": "system", 00:10:55.985 "dma_device_type": 1 00:10:55.985 }, 00:10:55.985 { 00:10:55.985 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:55.985 "dma_device_type": 2 00:10:55.985 }, 00:10:55.985 { 00:10:55.985 "dma_device_id": "system", 00:10:55.985 "dma_device_type": 1 00:10:55.985 }, 00:10:55.985 { 00:10:55.985 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:55.985 "dma_device_type": 2 00:10:55.985 }, 00:10:55.985 { 00:10:55.985 "dma_device_id": "system", 00:10:55.985 "dma_device_type": 1 00:10:55.985 }, 00:10:55.985 { 00:10:55.985 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:55.985 "dma_device_type": 2 00:10:55.985 } 00:10:55.985 ], 00:10:55.985 "driver_specific": { 00:10:55.985 "raid": { 00:10:55.985 "uuid": "665b16ec-9474-4950-a758-26cd1316d578", 00:10:55.985 "strip_size_kb": 0, 00:10:55.985 "state": "online", 00:10:55.985 "raid_level": "raid1", 00:10:55.985 "superblock": true, 00:10:55.985 "num_base_bdevs": 4, 00:10:55.985 "num_base_bdevs_discovered": 4, 00:10:55.985 "num_base_bdevs_operational": 4, 00:10:55.985 "base_bdevs_list": [ 00:10:55.985 { 00:10:55.985 "name": "pt1", 00:10:55.985 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:55.985 "is_configured": true, 00:10:55.985 "data_offset": 2048, 00:10:55.985 "data_size": 63488 00:10:55.985 }, 00:10:55.985 { 00:10:55.985 "name": "pt2", 00:10:55.985 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:55.985 "is_configured": true, 00:10:55.985 "data_offset": 2048, 00:10:55.985 "data_size": 63488 00:10:55.985 }, 00:10:55.985 { 00:10:55.985 "name": "pt3", 00:10:55.985 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:55.985 "is_configured": true, 00:10:55.985 "data_offset": 2048, 00:10:55.985 "data_size": 63488 00:10:55.985 }, 00:10:55.985 { 00:10:55.985 "name": "pt4", 00:10:55.985 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:55.985 "is_configured": true, 00:10:55.985 "data_offset": 2048, 00:10:55.985 "data_size": 63488 00:10:55.985 } 00:10:55.985 ] 00:10:55.985 } 00:10:55.985 } 00:10:55.985 }' 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:55.985 pt2 00:10:55.985 pt3 00:10:55.985 pt4' 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:55.985 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.245 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.245 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:56.245 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:56.245 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:56.245 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:56.245 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.245 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.245 [2024-12-05 19:01:13.580207] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:56.245 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.245 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 665b16ec-9474-4950-a758-26cd1316d578 '!=' 665b16ec-9474-4950-a758-26cd1316d578 ']' 00:10:56.245 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:10:56.245 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.246 [2024-12-05 19:01:13.611967] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:56.246 "name": "raid_bdev1", 00:10:56.246 "uuid": "665b16ec-9474-4950-a758-26cd1316d578", 00:10:56.246 "strip_size_kb": 0, 00:10:56.246 "state": "online", 00:10:56.246 "raid_level": "raid1", 00:10:56.246 "superblock": true, 00:10:56.246 "num_base_bdevs": 4, 00:10:56.246 "num_base_bdevs_discovered": 3, 00:10:56.246 "num_base_bdevs_operational": 3, 00:10:56.246 "base_bdevs_list": [ 00:10:56.246 { 00:10:56.246 "name": null, 00:10:56.246 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:56.246 "is_configured": false, 00:10:56.246 "data_offset": 0, 00:10:56.246 "data_size": 63488 00:10:56.246 }, 00:10:56.246 { 00:10:56.246 "name": "pt2", 00:10:56.246 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:56.246 "is_configured": true, 00:10:56.246 "data_offset": 2048, 00:10:56.246 "data_size": 63488 00:10:56.246 }, 00:10:56.246 { 00:10:56.246 "name": "pt3", 00:10:56.246 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:56.246 "is_configured": true, 00:10:56.246 "data_offset": 2048, 00:10:56.246 "data_size": 63488 00:10:56.246 }, 00:10:56.246 { 00:10:56.246 "name": "pt4", 00:10:56.246 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:56.246 "is_configured": true, 00:10:56.246 "data_offset": 2048, 00:10:56.246 "data_size": 63488 00:10:56.246 } 00:10:56.246 ] 00:10:56.246 }' 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:56.246 19:01:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.506 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:56.506 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.506 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.506 [2024-12-05 19:01:14.051326] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:56.506 [2024-12-05 19:01:14.051356] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:56.506 [2024-12-05 19:01:14.051444] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:56.506 [2024-12-05 19:01:14.051514] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:56.506 [2024-12-05 19:01:14.051524] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:10:56.506 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.506 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:10:56.506 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:56.506 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.506 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.765 [2024-12-05 19:01:14.131190] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:56.765 [2024-12-05 19:01:14.131261] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:56.765 [2024-12-05 19:01:14.131276] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:10:56.765 [2024-12-05 19:01:14.131286] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:56.765 [2024-12-05 19:01:14.133339] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:56.765 [2024-12-05 19:01:14.133380] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:56.765 [2024-12-05 19:01:14.133442] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:56.765 [2024-12-05 19:01:14.133477] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:56.765 pt2 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:56.765 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:56.766 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:56.766 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:56.766 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:56.766 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:56.766 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:56.766 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:56.766 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:56.766 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:56.766 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.766 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:56.766 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:56.766 "name": "raid_bdev1", 00:10:56.766 "uuid": "665b16ec-9474-4950-a758-26cd1316d578", 00:10:56.766 "strip_size_kb": 0, 00:10:56.766 "state": "configuring", 00:10:56.766 "raid_level": "raid1", 00:10:56.766 "superblock": true, 00:10:56.766 "num_base_bdevs": 4, 00:10:56.766 "num_base_bdevs_discovered": 1, 00:10:56.766 "num_base_bdevs_operational": 3, 00:10:56.766 "base_bdevs_list": [ 00:10:56.766 { 00:10:56.766 "name": null, 00:10:56.766 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:56.766 "is_configured": false, 00:10:56.766 "data_offset": 2048, 00:10:56.766 "data_size": 63488 00:10:56.766 }, 00:10:56.766 { 00:10:56.766 "name": "pt2", 00:10:56.766 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:56.766 "is_configured": true, 00:10:56.766 "data_offset": 2048, 00:10:56.766 "data_size": 63488 00:10:56.766 }, 00:10:56.766 { 00:10:56.766 "name": null, 00:10:56.766 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:56.766 "is_configured": false, 00:10:56.766 "data_offset": 2048, 00:10:56.766 "data_size": 63488 00:10:56.766 }, 00:10:56.766 { 00:10:56.766 "name": null, 00:10:56.766 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:56.766 "is_configured": false, 00:10:56.766 "data_offset": 2048, 00:10:56.766 "data_size": 63488 00:10:56.766 } 00:10:56.766 ] 00:10:56.766 }' 00:10:56.766 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:56.766 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.025 [2024-12-05 19:01:14.574439] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:57.025 [2024-12-05 19:01:14.574519] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:57.025 [2024-12-05 19:01:14.574538] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:10:57.025 [2024-12-05 19:01:14.574551] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:57.025 [2024-12-05 19:01:14.574914] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:57.025 [2024-12-05 19:01:14.574946] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:57.025 [2024-12-05 19:01:14.575010] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:57.025 [2024-12-05 19:01:14.575036] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:57.025 pt3 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:57.025 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:57.285 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.285 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.285 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.285 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:57.285 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.285 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:57.285 "name": "raid_bdev1", 00:10:57.285 "uuid": "665b16ec-9474-4950-a758-26cd1316d578", 00:10:57.285 "strip_size_kb": 0, 00:10:57.285 "state": "configuring", 00:10:57.285 "raid_level": "raid1", 00:10:57.285 "superblock": true, 00:10:57.285 "num_base_bdevs": 4, 00:10:57.285 "num_base_bdevs_discovered": 2, 00:10:57.286 "num_base_bdevs_operational": 3, 00:10:57.286 "base_bdevs_list": [ 00:10:57.286 { 00:10:57.286 "name": null, 00:10:57.286 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:57.286 "is_configured": false, 00:10:57.286 "data_offset": 2048, 00:10:57.286 "data_size": 63488 00:10:57.286 }, 00:10:57.286 { 00:10:57.286 "name": "pt2", 00:10:57.286 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:57.286 "is_configured": true, 00:10:57.286 "data_offset": 2048, 00:10:57.286 "data_size": 63488 00:10:57.286 }, 00:10:57.286 { 00:10:57.286 "name": "pt3", 00:10:57.286 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:57.286 "is_configured": true, 00:10:57.286 "data_offset": 2048, 00:10:57.286 "data_size": 63488 00:10:57.286 }, 00:10:57.286 { 00:10:57.286 "name": null, 00:10:57.286 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:57.286 "is_configured": false, 00:10:57.286 "data_offset": 2048, 00:10:57.286 "data_size": 63488 00:10:57.286 } 00:10:57.286 ] 00:10:57.286 }' 00:10:57.286 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:57.286 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.552 [2024-12-05 19:01:14.965741] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:57.552 [2024-12-05 19:01:14.965795] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:57.552 [2024-12-05 19:01:14.965812] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:10:57.552 [2024-12-05 19:01:14.965822] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:57.552 [2024-12-05 19:01:14.966188] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:57.552 [2024-12-05 19:01:14.966221] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:57.552 [2024-12-05 19:01:14.966286] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:57.552 [2024-12-05 19:01:14.966311] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:57.552 [2024-12-05 19:01:14.966404] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:57.552 [2024-12-05 19:01:14.966424] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:57.552 [2024-12-05 19:01:14.966685] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:10:57.552 [2024-12-05 19:01:14.966814] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:57.552 [2024-12-05 19:01:14.966833] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:10:57.552 [2024-12-05 19:01:14.966937] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:57.552 pt4 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.552 19:01:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:57.552 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:57.552 "name": "raid_bdev1", 00:10:57.552 "uuid": "665b16ec-9474-4950-a758-26cd1316d578", 00:10:57.552 "strip_size_kb": 0, 00:10:57.552 "state": "online", 00:10:57.552 "raid_level": "raid1", 00:10:57.552 "superblock": true, 00:10:57.552 "num_base_bdevs": 4, 00:10:57.552 "num_base_bdevs_discovered": 3, 00:10:57.552 "num_base_bdevs_operational": 3, 00:10:57.552 "base_bdevs_list": [ 00:10:57.552 { 00:10:57.552 "name": null, 00:10:57.552 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:57.552 "is_configured": false, 00:10:57.552 "data_offset": 2048, 00:10:57.552 "data_size": 63488 00:10:57.552 }, 00:10:57.552 { 00:10:57.552 "name": "pt2", 00:10:57.552 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:57.552 "is_configured": true, 00:10:57.552 "data_offset": 2048, 00:10:57.552 "data_size": 63488 00:10:57.552 }, 00:10:57.552 { 00:10:57.552 "name": "pt3", 00:10:57.552 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:57.552 "is_configured": true, 00:10:57.552 "data_offset": 2048, 00:10:57.552 "data_size": 63488 00:10:57.552 }, 00:10:57.552 { 00:10:57.552 "name": "pt4", 00:10:57.552 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:57.552 "is_configured": true, 00:10:57.552 "data_offset": 2048, 00:10:57.552 "data_size": 63488 00:10:57.552 } 00:10:57.552 ] 00:10:57.552 }' 00:10:57.552 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:57.552 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.128 [2024-12-05 19:01:15.412937] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:58.128 [2024-12-05 19:01:15.412969] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:58.128 [2024-12-05 19:01:15.413030] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:58.128 [2024-12-05 19:01:15.413099] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:58.128 [2024-12-05 19:01:15.413107] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.128 [2024-12-05 19:01:15.468859] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:58.128 [2024-12-05 19:01:15.468911] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:58.128 [2024-12-05 19:01:15.468944] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:10:58.128 [2024-12-05 19:01:15.468952] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:58.128 [2024-12-05 19:01:15.471009] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:58.128 [2024-12-05 19:01:15.471044] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:58.128 [2024-12-05 19:01:15.471106] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:58.128 [2024-12-05 19:01:15.471136] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:58.128 [2024-12-05 19:01:15.471246] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:10:58.128 [2024-12-05 19:01:15.471260] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:58.128 [2024-12-05 19:01:15.471276] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:10:58.128 [2024-12-05 19:01:15.471323] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:58.128 [2024-12-05 19:01:15.471415] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:58.128 pt1 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:58.128 "name": "raid_bdev1", 00:10:58.128 "uuid": "665b16ec-9474-4950-a758-26cd1316d578", 00:10:58.128 "strip_size_kb": 0, 00:10:58.128 "state": "configuring", 00:10:58.128 "raid_level": "raid1", 00:10:58.128 "superblock": true, 00:10:58.128 "num_base_bdevs": 4, 00:10:58.128 "num_base_bdevs_discovered": 2, 00:10:58.128 "num_base_bdevs_operational": 3, 00:10:58.128 "base_bdevs_list": [ 00:10:58.128 { 00:10:58.128 "name": null, 00:10:58.128 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:58.128 "is_configured": false, 00:10:58.128 "data_offset": 2048, 00:10:58.128 "data_size": 63488 00:10:58.128 }, 00:10:58.128 { 00:10:58.128 "name": "pt2", 00:10:58.128 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:58.128 "is_configured": true, 00:10:58.128 "data_offset": 2048, 00:10:58.128 "data_size": 63488 00:10:58.128 }, 00:10:58.128 { 00:10:58.128 "name": "pt3", 00:10:58.128 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:58.128 "is_configured": true, 00:10:58.128 "data_offset": 2048, 00:10:58.128 "data_size": 63488 00:10:58.128 }, 00:10:58.128 { 00:10:58.128 "name": null, 00:10:58.128 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:58.128 "is_configured": false, 00:10:58.128 "data_offset": 2048, 00:10:58.128 "data_size": 63488 00:10:58.128 } 00:10:58.128 ] 00:10:58.128 }' 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:58.128 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.386 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:10:58.386 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:10:58.386 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.386 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.386 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.644 [2024-12-05 19:01:15.964053] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:58.644 [2024-12-05 19:01:15.964107] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:58.644 [2024-12-05 19:01:15.964122] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:10:58.644 [2024-12-05 19:01:15.964132] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:58.644 [2024-12-05 19:01:15.964493] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:58.644 [2024-12-05 19:01:15.964526] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:58.644 [2024-12-05 19:01:15.964583] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:58.644 [2024-12-05 19:01:15.964606] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:58.644 [2024-12-05 19:01:15.964715] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:10:58.644 [2024-12-05 19:01:15.964727] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:58.644 [2024-12-05 19:01:15.964974] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:10:58.644 [2024-12-05 19:01:15.965106] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:10:58.644 [2024-12-05 19:01:15.965139] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:10:58.644 [2024-12-05 19:01:15.965242] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:58.644 pt4 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.644 19:01:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.644 19:01:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:58.644 "name": "raid_bdev1", 00:10:58.644 "uuid": "665b16ec-9474-4950-a758-26cd1316d578", 00:10:58.644 "strip_size_kb": 0, 00:10:58.644 "state": "online", 00:10:58.644 "raid_level": "raid1", 00:10:58.644 "superblock": true, 00:10:58.644 "num_base_bdevs": 4, 00:10:58.644 "num_base_bdevs_discovered": 3, 00:10:58.644 "num_base_bdevs_operational": 3, 00:10:58.644 "base_bdevs_list": [ 00:10:58.644 { 00:10:58.644 "name": null, 00:10:58.644 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:58.644 "is_configured": false, 00:10:58.644 "data_offset": 2048, 00:10:58.644 "data_size": 63488 00:10:58.644 }, 00:10:58.644 { 00:10:58.644 "name": "pt2", 00:10:58.644 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:58.644 "is_configured": true, 00:10:58.644 "data_offset": 2048, 00:10:58.644 "data_size": 63488 00:10:58.644 }, 00:10:58.644 { 00:10:58.644 "name": "pt3", 00:10:58.644 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:58.644 "is_configured": true, 00:10:58.644 "data_offset": 2048, 00:10:58.644 "data_size": 63488 00:10:58.644 }, 00:10:58.644 { 00:10:58.644 "name": "pt4", 00:10:58.644 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:58.644 "is_configured": true, 00:10:58.644 "data_offset": 2048, 00:10:58.645 "data_size": 63488 00:10:58.645 } 00:10:58.645 ] 00:10:58.645 }' 00:10:58.645 19:01:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:58.645 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.903 19:01:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:10:58.903 19:01:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:10:58.903 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.903 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.903 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:58.903 19:01:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:10:58.903 19:01:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:58.903 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:58.903 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.903 19:01:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:10:59.162 [2024-12-05 19:01:16.463519] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:59.162 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:59.162 19:01:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 665b16ec-9474-4950-a758-26cd1316d578 '!=' 665b16ec-9474-4950-a758-26cd1316d578 ']' 00:10:59.162 19:01:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 84116 00:10:59.162 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 84116 ']' 00:10:59.162 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@958 -- # kill -0 84116 00:10:59.162 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # uname 00:10:59.162 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:59.162 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84116 00:10:59.162 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:59.162 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:59.162 killing process with pid 84116 00:10:59.162 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84116' 00:10:59.162 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@973 -- # kill 84116 00:10:59.162 [2024-12-05 19:01:16.550451] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:59.162 [2024-12-05 19:01:16.550541] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:59.162 [2024-12-05 19:01:16.550618] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:59.162 [2024-12-05 19:01:16.550632] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:10:59.162 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@978 -- # wait 84116 00:10:59.162 [2024-12-05 19:01:16.592639] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:59.420 19:01:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:59.420 00:10:59.420 real 0m6.868s 00:10:59.420 user 0m11.565s 00:10:59.420 sys 0m1.433s 00:10:59.420 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:59.420 19:01:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.420 ************************************ 00:10:59.420 END TEST raid_superblock_test 00:10:59.420 ************************************ 00:10:59.420 19:01:16 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 4 read 00:10:59.420 19:01:16 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:10:59.420 19:01:16 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:59.420 19:01:16 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:59.420 ************************************ 00:10:59.420 START TEST raid_read_error_test 00:10:59.420 ************************************ 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 4 read 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:10:59.420 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:10:59.421 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:59.421 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.t7G1IXHY6d 00:10:59.421 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=84545 00:10:59.421 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:59.421 19:01:16 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 84545 00:10:59.421 19:01:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # '[' -z 84545 ']' 00:10:59.421 19:01:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:59.421 19:01:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:59.421 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:59.421 19:01:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:59.421 19:01:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:59.421 19:01:16 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.679 [2024-12-05 19:01:16.984410] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:10:59.679 [2024-12-05 19:01:16.984551] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid84545 ] 00:10:59.679 [2024-12-05 19:01:17.137501] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:59.679 [2024-12-05 19:01:17.163003] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:59.679 [2024-12-05 19:01:17.206100] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:59.679 [2024-12-05 19:01:17.206141] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:00.247 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:00.247 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@868 -- # return 0 00:11:00.247 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:00.247 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:00.247 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.247 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.508 BaseBdev1_malloc 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.508 true 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.508 [2024-12-05 19:01:17.834071] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:00.508 [2024-12-05 19:01:17.834149] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:00.508 [2024-12-05 19:01:17.834170] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:11:00.508 [2024-12-05 19:01:17.834179] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:00.508 [2024-12-05 19:01:17.836311] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:00.508 [2024-12-05 19:01:17.836350] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:00.508 BaseBdev1 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.508 BaseBdev2_malloc 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.508 true 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.508 [2024-12-05 19:01:17.874858] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:00.508 [2024-12-05 19:01:17.874923] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:00.508 [2024-12-05 19:01:17.874940] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:11:00.508 [2024-12-05 19:01:17.874956] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:00.508 [2024-12-05 19:01:17.877039] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:00.508 [2024-12-05 19:01:17.877078] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:00.508 BaseBdev2 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.508 BaseBdev3_malloc 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.508 true 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.508 [2024-12-05 19:01:17.915461] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:00.508 [2024-12-05 19:01:17.915525] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:00.508 [2024-12-05 19:01:17.915543] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:11:00.508 [2024-12-05 19:01:17.915552] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:00.508 [2024-12-05 19:01:17.917617] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:00.508 [2024-12-05 19:01:17.917653] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:00.508 BaseBdev3 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.508 BaseBdev4_malloc 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.508 true 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.508 [2024-12-05 19:01:17.973654] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:00.508 [2024-12-05 19:01:17.973722] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:00.508 [2024-12-05 19:01:17.973747] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:00.508 [2024-12-05 19:01:17.973758] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:00.508 [2024-12-05 19:01:17.976106] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:00.508 [2024-12-05 19:01:17.976149] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:00.508 BaseBdev4 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.508 [2024-12-05 19:01:17.985659] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:00.508 [2024-12-05 19:01:17.987530] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:00.508 [2024-12-05 19:01:17.987623] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:00.508 [2024-12-05 19:01:17.987673] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:00.508 [2024-12-05 19:01:17.987906] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:11:00.508 [2024-12-05 19:01:17.987927] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:00.508 [2024-12-05 19:01:17.988186] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:11:00.508 [2024-12-05 19:01:17.988335] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:11:00.508 [2024-12-05 19:01:17.988355] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:11:00.508 [2024-12-05 19:01:17.988478] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.508 19:01:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:00.508 19:01:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:00.508 19:01:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:00.508 "name": "raid_bdev1", 00:11:00.508 "uuid": "7b3bbe86-9ec1-42c7-baab-08d0ff988130", 00:11:00.509 "strip_size_kb": 0, 00:11:00.509 "state": "online", 00:11:00.509 "raid_level": "raid1", 00:11:00.509 "superblock": true, 00:11:00.509 "num_base_bdevs": 4, 00:11:00.509 "num_base_bdevs_discovered": 4, 00:11:00.509 "num_base_bdevs_operational": 4, 00:11:00.509 "base_bdevs_list": [ 00:11:00.509 { 00:11:00.509 "name": "BaseBdev1", 00:11:00.509 "uuid": "d3ddb44f-82aa-5a81-a903-d07fb6796e36", 00:11:00.509 "is_configured": true, 00:11:00.509 "data_offset": 2048, 00:11:00.509 "data_size": 63488 00:11:00.509 }, 00:11:00.509 { 00:11:00.509 "name": "BaseBdev2", 00:11:00.509 "uuid": "7dc7cee4-205f-532d-a1ee-e1cc03223228", 00:11:00.509 "is_configured": true, 00:11:00.509 "data_offset": 2048, 00:11:00.509 "data_size": 63488 00:11:00.509 }, 00:11:00.509 { 00:11:00.509 "name": "BaseBdev3", 00:11:00.509 "uuid": "41714c7f-f7e7-544b-8c6d-1a3270e44b46", 00:11:00.509 "is_configured": true, 00:11:00.509 "data_offset": 2048, 00:11:00.509 "data_size": 63488 00:11:00.509 }, 00:11:00.509 { 00:11:00.509 "name": "BaseBdev4", 00:11:00.509 "uuid": "3e894935-67c3-5a51-a18a-e39123bb29c0", 00:11:00.509 "is_configured": true, 00:11:00.509 "data_offset": 2048, 00:11:00.509 "data_size": 63488 00:11:00.509 } 00:11:00.509 ] 00:11:00.509 }' 00:11:00.509 19:01:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:00.509 19:01:18 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.076 19:01:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:01.076 19:01:18 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:01.076 [2024-12-05 19:01:18.473213] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:02.010 "name": "raid_bdev1", 00:11:02.010 "uuid": "7b3bbe86-9ec1-42c7-baab-08d0ff988130", 00:11:02.010 "strip_size_kb": 0, 00:11:02.010 "state": "online", 00:11:02.010 "raid_level": "raid1", 00:11:02.010 "superblock": true, 00:11:02.010 "num_base_bdevs": 4, 00:11:02.010 "num_base_bdevs_discovered": 4, 00:11:02.010 "num_base_bdevs_operational": 4, 00:11:02.010 "base_bdevs_list": [ 00:11:02.010 { 00:11:02.010 "name": "BaseBdev1", 00:11:02.010 "uuid": "d3ddb44f-82aa-5a81-a903-d07fb6796e36", 00:11:02.010 "is_configured": true, 00:11:02.010 "data_offset": 2048, 00:11:02.010 "data_size": 63488 00:11:02.010 }, 00:11:02.010 { 00:11:02.010 "name": "BaseBdev2", 00:11:02.010 "uuid": "7dc7cee4-205f-532d-a1ee-e1cc03223228", 00:11:02.010 "is_configured": true, 00:11:02.010 "data_offset": 2048, 00:11:02.010 "data_size": 63488 00:11:02.010 }, 00:11:02.010 { 00:11:02.010 "name": "BaseBdev3", 00:11:02.010 "uuid": "41714c7f-f7e7-544b-8c6d-1a3270e44b46", 00:11:02.010 "is_configured": true, 00:11:02.010 "data_offset": 2048, 00:11:02.010 "data_size": 63488 00:11:02.010 }, 00:11:02.010 { 00:11:02.010 "name": "BaseBdev4", 00:11:02.010 "uuid": "3e894935-67c3-5a51-a18a-e39123bb29c0", 00:11:02.010 "is_configured": true, 00:11:02.010 "data_offset": 2048, 00:11:02.010 "data_size": 63488 00:11:02.010 } 00:11:02.010 ] 00:11:02.010 }' 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:02.010 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.577 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:02.577 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:02.577 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.577 [2024-12-05 19:01:19.849753] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:02.577 [2024-12-05 19:01:19.849788] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:02.577 [2024-12-05 19:01:19.852352] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:02.577 [2024-12-05 19:01:19.852428] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:02.577 [2024-12-05 19:01:19.852548] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:02.577 [2024-12-05 19:01:19.852570] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:11:02.577 { 00:11:02.577 "results": [ 00:11:02.577 { 00:11:02.577 "job": "raid_bdev1", 00:11:02.577 "core_mask": "0x1", 00:11:02.577 "workload": "randrw", 00:11:02.577 "percentage": 50, 00:11:02.577 "status": "finished", 00:11:02.577 "queue_depth": 1, 00:11:02.577 "io_size": 131072, 00:11:02.577 "runtime": 1.377418, 00:11:02.577 "iops": 11657.318257783767, 00:11:02.577 "mibps": 1457.1647822229709, 00:11:02.577 "io_failed": 0, 00:11:02.577 "io_timeout": 0, 00:11:02.577 "avg_latency_us": 83.25385008048565, 00:11:02.577 "min_latency_us": 23.14061135371179, 00:11:02.577 "max_latency_us": 1438.071615720524 00:11:02.577 } 00:11:02.577 ], 00:11:02.577 "core_count": 1 00:11:02.577 } 00:11:02.577 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:02.577 19:01:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 84545 00:11:02.577 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # '[' -z 84545 ']' 00:11:02.577 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@958 -- # kill -0 84545 00:11:02.577 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # uname 00:11:02.577 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:02.577 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84545 00:11:02.577 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:02.577 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:02.577 killing process with pid 84545 00:11:02.577 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84545' 00:11:02.577 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@973 -- # kill 84545 00:11:02.577 [2024-12-05 19:01:19.893382] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:02.577 19:01:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@978 -- # wait 84545 00:11:02.577 [2024-12-05 19:01:19.927453] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:02.577 19:01:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.t7G1IXHY6d 00:11:02.577 19:01:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:02.577 19:01:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:02.836 19:01:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:11:02.836 19:01:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:11:02.836 19:01:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:02.836 19:01:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:11:02.836 19:01:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:11:02.836 00:11:02.836 real 0m3.260s 00:11:02.836 user 0m4.076s 00:11:02.836 sys 0m0.519s 00:11:02.836 19:01:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:02.836 19:01:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.836 ************************************ 00:11:02.836 END TEST raid_read_error_test 00:11:02.836 ************************************ 00:11:02.836 19:01:20 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 4 write 00:11:02.836 19:01:20 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:11:02.836 19:01:20 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:02.836 19:01:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:02.836 ************************************ 00:11:02.836 START TEST raid_write_error_test 00:11:02.836 ************************************ 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1129 -- # raid_io_error_test raid1 4 write 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.qf01fbTTJB 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=84657 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 84657 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # '[' -z 84657 ']' 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:02.836 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:02.836 19:01:20 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.836 [2024-12-05 19:01:20.317578] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:11:02.836 [2024-12-05 19:01:20.317724] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid84657 ] 00:11:03.094 [2024-12-05 19:01:20.471721] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:03.094 [2024-12-05 19:01:20.496604] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:03.094 [2024-12-05 19:01:20.539545] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:03.094 [2024-12-05 19:01:20.539589] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:03.660 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:03.660 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@868 -- # return 0 00:11:03.660 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:03.660 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:03.660 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.660 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.660 BaseBdev1_malloc 00:11:03.660 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.660 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:03.660 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.660 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.660 true 00:11:03.660 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.660 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:03.660 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.660 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.660 [2024-12-05 19:01:21.171278] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:03.661 [2024-12-05 19:01:21.171354] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:03.661 [2024-12-05 19:01:21.171376] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:11:03.661 [2024-12-05 19:01:21.171385] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:03.661 [2024-12-05 19:01:21.173467] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:03.661 [2024-12-05 19:01:21.173509] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:03.661 BaseBdev1 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.661 BaseBdev2_malloc 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.661 true 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.661 [2024-12-05 19:01:21.211805] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:03.661 [2024-12-05 19:01:21.211863] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:03.661 [2024-12-05 19:01:21.211897] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:11:03.661 [2024-12-05 19:01:21.211914] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:03.661 [2024-12-05 19:01:21.213974] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:03.661 [2024-12-05 19:01:21.214011] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:03.661 BaseBdev2 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.661 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.919 BaseBdev3_malloc 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.919 true 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.919 [2024-12-05 19:01:21.252385] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:03.919 [2024-12-05 19:01:21.252433] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:03.919 [2024-12-05 19:01:21.252451] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:11:03.919 [2024-12-05 19:01:21.252460] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:03.919 [2024-12-05 19:01:21.254456] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:03.919 [2024-12-05 19:01:21.254491] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:03.919 BaseBdev3 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.919 BaseBdev4_malloc 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.919 true 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.919 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.919 [2024-12-05 19:01:21.302427] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:03.919 [2024-12-05 19:01:21.302491] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:03.919 [2024-12-05 19:01:21.302512] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:03.920 [2024-12-05 19:01:21.302521] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:03.920 [2024-12-05 19:01:21.304506] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:03.920 [2024-12-05 19:01:21.304545] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:03.920 BaseBdev4 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.920 [2024-12-05 19:01:21.314449] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:03.920 [2024-12-05 19:01:21.316259] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:03.920 [2024-12-05 19:01:21.316339] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:03.920 [2024-12-05 19:01:21.316388] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:03.920 [2024-12-05 19:01:21.316593] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:11:03.920 [2024-12-05 19:01:21.316613] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:03.920 [2024-12-05 19:01:21.316870] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002ef0 00:11:03.920 [2024-12-05 19:01:21.317022] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:11:03.920 [2024-12-05 19:01:21.317042] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:11:03.920 [2024-12-05 19:01:21.317176] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:03.920 "name": "raid_bdev1", 00:11:03.920 "uuid": "d72370a9-2042-46d0-b6e1-27ad81a95a2f", 00:11:03.920 "strip_size_kb": 0, 00:11:03.920 "state": "online", 00:11:03.920 "raid_level": "raid1", 00:11:03.920 "superblock": true, 00:11:03.920 "num_base_bdevs": 4, 00:11:03.920 "num_base_bdevs_discovered": 4, 00:11:03.920 "num_base_bdevs_operational": 4, 00:11:03.920 "base_bdevs_list": [ 00:11:03.920 { 00:11:03.920 "name": "BaseBdev1", 00:11:03.920 "uuid": "88ba7aec-e21a-5a34-88eb-c5469e303978", 00:11:03.920 "is_configured": true, 00:11:03.920 "data_offset": 2048, 00:11:03.920 "data_size": 63488 00:11:03.920 }, 00:11:03.920 { 00:11:03.920 "name": "BaseBdev2", 00:11:03.920 "uuid": "c0dea25f-d5c6-51da-9e56-7b769b81d4a6", 00:11:03.920 "is_configured": true, 00:11:03.920 "data_offset": 2048, 00:11:03.920 "data_size": 63488 00:11:03.920 }, 00:11:03.920 { 00:11:03.920 "name": "BaseBdev3", 00:11:03.920 "uuid": "6a890ea5-9b8f-5692-9c71-f73c097c3361", 00:11:03.920 "is_configured": true, 00:11:03.920 "data_offset": 2048, 00:11:03.920 "data_size": 63488 00:11:03.920 }, 00:11:03.920 { 00:11:03.920 "name": "BaseBdev4", 00:11:03.920 "uuid": "c92ec682-c5cd-56c6-b6e4-eeadaa5a5605", 00:11:03.920 "is_configured": true, 00:11:03.920 "data_offset": 2048, 00:11:03.920 "data_size": 63488 00:11:03.920 } 00:11:03.920 ] 00:11:03.920 }' 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:03.920 19:01:21 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.178 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:04.178 19:01:21 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:04.437 [2024-12-05 19:01:21.801986] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000003090 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.374 [2024-12-05 19:01:22.745100] bdev_raid.c:2276:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:11:05.374 [2024-12-05 19:01:22.745158] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:05.374 [2024-12-05 19:01:22.745405] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000003090 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.374 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:05.374 "name": "raid_bdev1", 00:11:05.374 "uuid": "d72370a9-2042-46d0-b6e1-27ad81a95a2f", 00:11:05.374 "strip_size_kb": 0, 00:11:05.374 "state": "online", 00:11:05.374 "raid_level": "raid1", 00:11:05.374 "superblock": true, 00:11:05.374 "num_base_bdevs": 4, 00:11:05.374 "num_base_bdevs_discovered": 3, 00:11:05.374 "num_base_bdevs_operational": 3, 00:11:05.374 "base_bdevs_list": [ 00:11:05.374 { 00:11:05.375 "name": null, 00:11:05.375 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:05.375 "is_configured": false, 00:11:05.375 "data_offset": 0, 00:11:05.375 "data_size": 63488 00:11:05.375 }, 00:11:05.375 { 00:11:05.375 "name": "BaseBdev2", 00:11:05.375 "uuid": "c0dea25f-d5c6-51da-9e56-7b769b81d4a6", 00:11:05.375 "is_configured": true, 00:11:05.375 "data_offset": 2048, 00:11:05.375 "data_size": 63488 00:11:05.375 }, 00:11:05.375 { 00:11:05.375 "name": "BaseBdev3", 00:11:05.375 "uuid": "6a890ea5-9b8f-5692-9c71-f73c097c3361", 00:11:05.375 "is_configured": true, 00:11:05.375 "data_offset": 2048, 00:11:05.375 "data_size": 63488 00:11:05.375 }, 00:11:05.375 { 00:11:05.375 "name": "BaseBdev4", 00:11:05.375 "uuid": "c92ec682-c5cd-56c6-b6e4-eeadaa5a5605", 00:11:05.375 "is_configured": true, 00:11:05.375 "data_offset": 2048, 00:11:05.375 "data_size": 63488 00:11:05.375 } 00:11:05.375 ] 00:11:05.375 }' 00:11:05.375 19:01:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:05.375 19:01:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.634 19:01:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:05.634 19:01:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:05.634 19:01:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.634 [2024-12-05 19:01:23.188088] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:05.634 [2024-12-05 19:01:23.188129] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:05.634 [2024-12-05 19:01:23.190729] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:05.634 [2024-12-05 19:01:23.190788] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:05.634 [2024-12-05 19:01:23.190899] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:05.634 [2024-12-05 19:01:23.190924] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:11:05.894 { 00:11:05.894 "results": [ 00:11:05.894 { 00:11:05.894 "job": "raid_bdev1", 00:11:05.894 "core_mask": "0x1", 00:11:05.894 "workload": "randrw", 00:11:05.894 "percentage": 50, 00:11:05.894 "status": "finished", 00:11:05.894 "queue_depth": 1, 00:11:05.894 "io_size": 131072, 00:11:05.894 "runtime": 1.387128, 00:11:05.894 "iops": 12792.618994065437, 00:11:05.894 "mibps": 1599.0773742581796, 00:11:05.894 "io_failed": 0, 00:11:05.894 "io_timeout": 0, 00:11:05.894 "avg_latency_us": 75.64029358168425, 00:11:05.894 "min_latency_us": 22.805240174672488, 00:11:05.894 "max_latency_us": 1337.907423580786 00:11:05.894 } 00:11:05.894 ], 00:11:05.894 "core_count": 1 00:11:05.894 } 00:11:05.894 19:01:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:05.894 19:01:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 84657 00:11:05.894 19:01:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # '[' -z 84657 ']' 00:11:05.894 19:01:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@958 -- # kill -0 84657 00:11:05.894 19:01:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # uname 00:11:05.894 19:01:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:05.894 19:01:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84657 00:11:05.894 19:01:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:05.894 19:01:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:05.894 killing process with pid 84657 00:11:05.894 19:01:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84657' 00:11:05.894 19:01:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@973 -- # kill 84657 00:11:05.894 [2024-12-05 19:01:23.240011] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:05.894 19:01:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@978 -- # wait 84657 00:11:05.894 [2024-12-05 19:01:23.274411] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:06.155 19:01:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.qf01fbTTJB 00:11:06.155 19:01:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:06.155 19:01:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:06.155 19:01:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:11:06.155 19:01:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:11:06.155 19:01:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:06.155 19:01:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:11:06.155 19:01:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:11:06.155 00:11:06.155 real 0m3.266s 00:11:06.155 user 0m4.075s 00:11:06.155 sys 0m0.542s 00:11:06.155 19:01:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:06.155 19:01:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.155 ************************************ 00:11:06.155 END TEST raid_write_error_test 00:11:06.155 ************************************ 00:11:06.155 19:01:23 bdev_raid -- bdev/bdev_raid.sh@976 -- # '[' true = true ']' 00:11:06.155 19:01:23 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:11:06.155 19:01:23 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 2 false false true 00:11:06.155 19:01:23 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:11:06.155 19:01:23 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:06.155 19:01:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:06.155 ************************************ 00:11:06.155 START TEST raid_rebuild_test 00:11:06.155 ************************************ 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 false false true 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=84766 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 84766 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 84766 ']' 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:06.155 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:06.155 19:01:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.155 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:06.155 Zero copy mechanism will not be used. 00:11:06.155 [2024-12-05 19:01:23.652607] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:11:06.155 [2024-12-05 19:01:23.652750] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid84766 ] 00:11:06.415 [2024-12-05 19:01:23.806618] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:06.415 [2024-12-05 19:01:23.831348] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:06.415 [2024-12-05 19:01:23.874358] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:06.415 [2024-12-05 19:01:23.874402] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:06.985 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:06.985 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:11:06.985 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:06.985 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:06.985 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.985 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.985 BaseBdev1_malloc 00:11:06.985 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.985 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:06.985 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.985 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.985 [2024-12-05 19:01:24.490354] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:06.985 [2024-12-05 19:01:24.490433] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:06.985 [2024-12-05 19:01:24.490458] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:06.985 [2024-12-05 19:01:24.490470] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:06.985 [2024-12-05 19:01:24.492602] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:06.985 [2024-12-05 19:01:24.492644] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:06.985 BaseBdev1 00:11:06.985 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.985 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:06.986 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:06.986 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.986 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.986 BaseBdev2_malloc 00:11:06.986 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.986 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:06.986 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.986 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.986 [2024-12-05 19:01:24.518960] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:06.986 [2024-12-05 19:01:24.519030] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:06.986 [2024-12-05 19:01:24.519053] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:06.986 [2024-12-05 19:01:24.519062] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:06.986 [2024-12-05 19:01:24.521130] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:06.986 [2024-12-05 19:01:24.521174] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:06.986 BaseBdev2 00:11:06.986 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.986 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:06.986 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.986 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.986 spare_malloc 00:11:06.986 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:06.986 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:06.986 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:06.986 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.247 spare_delay 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.247 [2024-12-05 19:01:24.559530] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:07.247 [2024-12-05 19:01:24.559593] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:07.247 [2024-12-05 19:01:24.559611] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:07.247 [2024-12-05 19:01:24.559619] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:07.247 [2024-12-05 19:01:24.561738] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:07.247 [2024-12-05 19:01:24.561774] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:07.247 spare 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.247 [2024-12-05 19:01:24.571540] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:07.247 [2024-12-05 19:01:24.573423] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:07.247 [2024-12-05 19:01:24.573514] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:07.247 [2024-12-05 19:01:24.573524] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:07.247 [2024-12-05 19:01:24.573816] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:11:07.247 [2024-12-05 19:01:24.573964] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:07.247 [2024-12-05 19:01:24.573985] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:07.247 [2024-12-05 19:01:24.574092] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:07.247 "name": "raid_bdev1", 00:11:07.247 "uuid": "977c5b1f-d8aa-4a24-ba19-a43b28b8bc11", 00:11:07.247 "strip_size_kb": 0, 00:11:07.247 "state": "online", 00:11:07.247 "raid_level": "raid1", 00:11:07.247 "superblock": false, 00:11:07.247 "num_base_bdevs": 2, 00:11:07.247 "num_base_bdevs_discovered": 2, 00:11:07.247 "num_base_bdevs_operational": 2, 00:11:07.247 "base_bdevs_list": [ 00:11:07.247 { 00:11:07.247 "name": "BaseBdev1", 00:11:07.247 "uuid": "595114cf-e8b5-5bcb-9ad6-32263e1ea4ea", 00:11:07.247 "is_configured": true, 00:11:07.247 "data_offset": 0, 00:11:07.247 "data_size": 65536 00:11:07.247 }, 00:11:07.247 { 00:11:07.247 "name": "BaseBdev2", 00:11:07.247 "uuid": "d32a3e77-b327-5474-a9a6-73d122557ebe", 00:11:07.247 "is_configured": true, 00:11:07.247 "data_offset": 0, 00:11:07.247 "data_size": 65536 00:11:07.247 } 00:11:07.247 ] 00:11:07.247 }' 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:07.247 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.508 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:07.508 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.508 19:01:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.508 19:01:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:07.508 [2024-12-05 19:01:24.999047] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:07.508 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.508 19:01:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:11:07.508 19:01:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.508 19:01:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:07.508 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.508 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.508 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.765 19:01:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:11:07.765 19:01:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:11:07.765 19:01:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:11:07.765 19:01:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:11:07.765 19:01:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:11:07.765 19:01:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:07.765 19:01:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:11:07.765 19:01:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:07.765 19:01:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:07.765 19:01:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:07.765 19:01:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:11:07.765 19:01:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:07.765 19:01:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:07.765 19:01:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:11:07.765 [2024-12-05 19:01:25.278414] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:11:07.765 /dev/nbd0 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:08.023 1+0 records in 00:11:08.023 1+0 records out 00:11:08.023 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00038124 s, 10.7 MB/s 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:11:08.023 19:01:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:11:12.214 65536+0 records in 00:11:12.214 65536+0 records out 00:11:12.214 33554432 bytes (34 MB, 32 MiB) copied, 3.51486 s, 9.5 MB/s 00:11:12.214 19:01:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:12.214 19:01:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:12.214 19:01:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:12.214 19:01:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:12.214 19:01:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:11:12.214 19:01:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:12.214 19:01:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:12.214 [2024-12-05 19:01:29.089753] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.214 [2024-12-05 19:01:29.106783] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:12.214 "name": "raid_bdev1", 00:11:12.214 "uuid": "977c5b1f-d8aa-4a24-ba19-a43b28b8bc11", 00:11:12.214 "strip_size_kb": 0, 00:11:12.214 "state": "online", 00:11:12.214 "raid_level": "raid1", 00:11:12.214 "superblock": false, 00:11:12.214 "num_base_bdevs": 2, 00:11:12.214 "num_base_bdevs_discovered": 1, 00:11:12.214 "num_base_bdevs_operational": 1, 00:11:12.214 "base_bdevs_list": [ 00:11:12.214 { 00:11:12.214 "name": null, 00:11:12.214 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:12.214 "is_configured": false, 00:11:12.214 "data_offset": 0, 00:11:12.214 "data_size": 65536 00:11:12.214 }, 00:11:12.214 { 00:11:12.214 "name": "BaseBdev2", 00:11:12.214 "uuid": "d32a3e77-b327-5474-a9a6-73d122557ebe", 00:11:12.214 "is_configured": true, 00:11:12.214 "data_offset": 0, 00:11:12.214 "data_size": 65536 00:11:12.214 } 00:11:12.214 ] 00:11:12.214 }' 00:11:12.214 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:12.215 19:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.215 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:12.215 19:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:12.215 19:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:12.215 [2024-12-05 19:01:29.534024] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:12.215 [2024-12-05 19:01:29.539170] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d06220 00:11:12.215 19:01:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:12.215 19:01:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:12.215 [2024-12-05 19:01:29.541118] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:13.172 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:13.172 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:13.172 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:13.172 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:13.172 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:13.173 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:13.173 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:13.173 19:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.173 19:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.173 19:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.173 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:13.173 "name": "raid_bdev1", 00:11:13.173 "uuid": "977c5b1f-d8aa-4a24-ba19-a43b28b8bc11", 00:11:13.173 "strip_size_kb": 0, 00:11:13.173 "state": "online", 00:11:13.173 "raid_level": "raid1", 00:11:13.173 "superblock": false, 00:11:13.173 "num_base_bdevs": 2, 00:11:13.173 "num_base_bdevs_discovered": 2, 00:11:13.173 "num_base_bdevs_operational": 2, 00:11:13.173 "process": { 00:11:13.173 "type": "rebuild", 00:11:13.173 "target": "spare", 00:11:13.173 "progress": { 00:11:13.173 "blocks": 20480, 00:11:13.173 "percent": 31 00:11:13.173 } 00:11:13.173 }, 00:11:13.173 "base_bdevs_list": [ 00:11:13.173 { 00:11:13.173 "name": "spare", 00:11:13.173 "uuid": "717f5d56-410e-5dc7-8c1f-3e01f30852a9", 00:11:13.173 "is_configured": true, 00:11:13.173 "data_offset": 0, 00:11:13.173 "data_size": 65536 00:11:13.173 }, 00:11:13.173 { 00:11:13.173 "name": "BaseBdev2", 00:11:13.173 "uuid": "d32a3e77-b327-5474-a9a6-73d122557ebe", 00:11:13.173 "is_configured": true, 00:11:13.173 "data_offset": 0, 00:11:13.173 "data_size": 65536 00:11:13.173 } 00:11:13.173 ] 00:11:13.173 }' 00:11:13.173 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:13.173 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:13.173 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:13.173 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:13.173 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:13.173 19:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.173 19:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.173 [2024-12-05 19:01:30.685219] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:13.469 [2024-12-05 19:01:30.745624] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:13.469 [2024-12-05 19:01:30.745711] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:13.469 [2024-12-05 19:01:30.745731] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:13.469 [2024-12-05 19:01:30.745739] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:13.469 "name": "raid_bdev1", 00:11:13.469 "uuid": "977c5b1f-d8aa-4a24-ba19-a43b28b8bc11", 00:11:13.469 "strip_size_kb": 0, 00:11:13.469 "state": "online", 00:11:13.469 "raid_level": "raid1", 00:11:13.469 "superblock": false, 00:11:13.469 "num_base_bdevs": 2, 00:11:13.469 "num_base_bdevs_discovered": 1, 00:11:13.469 "num_base_bdevs_operational": 1, 00:11:13.469 "base_bdevs_list": [ 00:11:13.469 { 00:11:13.469 "name": null, 00:11:13.469 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:13.469 "is_configured": false, 00:11:13.469 "data_offset": 0, 00:11:13.469 "data_size": 65536 00:11:13.469 }, 00:11:13.469 { 00:11:13.469 "name": "BaseBdev2", 00:11:13.469 "uuid": "d32a3e77-b327-5474-a9a6-73d122557ebe", 00:11:13.469 "is_configured": true, 00:11:13.469 "data_offset": 0, 00:11:13.469 "data_size": 65536 00:11:13.469 } 00:11:13.469 ] 00:11:13.469 }' 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:13.469 19:01:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:13.745 "name": "raid_bdev1", 00:11:13.745 "uuid": "977c5b1f-d8aa-4a24-ba19-a43b28b8bc11", 00:11:13.745 "strip_size_kb": 0, 00:11:13.745 "state": "online", 00:11:13.745 "raid_level": "raid1", 00:11:13.745 "superblock": false, 00:11:13.745 "num_base_bdevs": 2, 00:11:13.745 "num_base_bdevs_discovered": 1, 00:11:13.745 "num_base_bdevs_operational": 1, 00:11:13.745 "base_bdevs_list": [ 00:11:13.745 { 00:11:13.745 "name": null, 00:11:13.745 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:13.745 "is_configured": false, 00:11:13.745 "data_offset": 0, 00:11:13.745 "data_size": 65536 00:11:13.745 }, 00:11:13.745 { 00:11:13.745 "name": "BaseBdev2", 00:11:13.745 "uuid": "d32a3e77-b327-5474-a9a6-73d122557ebe", 00:11:13.745 "is_configured": true, 00:11:13.745 "data_offset": 0, 00:11:13.745 "data_size": 65536 00:11:13.745 } 00:11:13.745 ] 00:11:13.745 }' 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:13.745 [2024-12-05 19:01:31.297481] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:13.745 [2024-12-05 19:01:31.302312] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d062f0 00:11:13.745 19:01:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:14.005 19:01:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:14.005 [2024-12-05 19:01:31.304284] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:14.943 "name": "raid_bdev1", 00:11:14.943 "uuid": "977c5b1f-d8aa-4a24-ba19-a43b28b8bc11", 00:11:14.943 "strip_size_kb": 0, 00:11:14.943 "state": "online", 00:11:14.943 "raid_level": "raid1", 00:11:14.943 "superblock": false, 00:11:14.943 "num_base_bdevs": 2, 00:11:14.943 "num_base_bdevs_discovered": 2, 00:11:14.943 "num_base_bdevs_operational": 2, 00:11:14.943 "process": { 00:11:14.943 "type": "rebuild", 00:11:14.943 "target": "spare", 00:11:14.943 "progress": { 00:11:14.943 "blocks": 20480, 00:11:14.943 "percent": 31 00:11:14.943 } 00:11:14.943 }, 00:11:14.943 "base_bdevs_list": [ 00:11:14.943 { 00:11:14.943 "name": "spare", 00:11:14.943 "uuid": "717f5d56-410e-5dc7-8c1f-3e01f30852a9", 00:11:14.943 "is_configured": true, 00:11:14.943 "data_offset": 0, 00:11:14.943 "data_size": 65536 00:11:14.943 }, 00:11:14.943 { 00:11:14.943 "name": "BaseBdev2", 00:11:14.943 "uuid": "d32a3e77-b327-5474-a9a6-73d122557ebe", 00:11:14.943 "is_configured": true, 00:11:14.943 "data_offset": 0, 00:11:14.943 "data_size": 65536 00:11:14.943 } 00:11:14.943 ] 00:11:14.943 }' 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=282 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:14.943 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:14.943 "name": "raid_bdev1", 00:11:14.943 "uuid": "977c5b1f-d8aa-4a24-ba19-a43b28b8bc11", 00:11:14.943 "strip_size_kb": 0, 00:11:14.943 "state": "online", 00:11:14.944 "raid_level": "raid1", 00:11:14.944 "superblock": false, 00:11:14.944 "num_base_bdevs": 2, 00:11:14.944 "num_base_bdevs_discovered": 2, 00:11:14.944 "num_base_bdevs_operational": 2, 00:11:14.944 "process": { 00:11:14.944 "type": "rebuild", 00:11:14.944 "target": "spare", 00:11:14.944 "progress": { 00:11:14.944 "blocks": 22528, 00:11:14.944 "percent": 34 00:11:14.944 } 00:11:14.944 }, 00:11:14.944 "base_bdevs_list": [ 00:11:14.944 { 00:11:14.944 "name": "spare", 00:11:14.944 "uuid": "717f5d56-410e-5dc7-8c1f-3e01f30852a9", 00:11:14.944 "is_configured": true, 00:11:14.944 "data_offset": 0, 00:11:14.944 "data_size": 65536 00:11:14.944 }, 00:11:14.944 { 00:11:14.944 "name": "BaseBdev2", 00:11:14.944 "uuid": "d32a3e77-b327-5474-a9a6-73d122557ebe", 00:11:14.944 "is_configured": true, 00:11:14.944 "data_offset": 0, 00:11:14.944 "data_size": 65536 00:11:14.944 } 00:11:14.944 ] 00:11:14.944 }' 00:11:14.944 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:15.203 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:15.203 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:15.203 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:15.203 19:01:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:16.142 "name": "raid_bdev1", 00:11:16.142 "uuid": "977c5b1f-d8aa-4a24-ba19-a43b28b8bc11", 00:11:16.142 "strip_size_kb": 0, 00:11:16.142 "state": "online", 00:11:16.142 "raid_level": "raid1", 00:11:16.142 "superblock": false, 00:11:16.142 "num_base_bdevs": 2, 00:11:16.142 "num_base_bdevs_discovered": 2, 00:11:16.142 "num_base_bdevs_operational": 2, 00:11:16.142 "process": { 00:11:16.142 "type": "rebuild", 00:11:16.142 "target": "spare", 00:11:16.142 "progress": { 00:11:16.142 "blocks": 45056, 00:11:16.142 "percent": 68 00:11:16.142 } 00:11:16.142 }, 00:11:16.142 "base_bdevs_list": [ 00:11:16.142 { 00:11:16.142 "name": "spare", 00:11:16.142 "uuid": "717f5d56-410e-5dc7-8c1f-3e01f30852a9", 00:11:16.142 "is_configured": true, 00:11:16.142 "data_offset": 0, 00:11:16.142 "data_size": 65536 00:11:16.142 }, 00:11:16.142 { 00:11:16.142 "name": "BaseBdev2", 00:11:16.142 "uuid": "d32a3e77-b327-5474-a9a6-73d122557ebe", 00:11:16.142 "is_configured": true, 00:11:16.142 "data_offset": 0, 00:11:16.142 "data_size": 65536 00:11:16.142 } 00:11:16.142 ] 00:11:16.142 }' 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:16.142 19:01:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:17.080 [2024-12-05 19:01:34.514814] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:17.080 [2024-12-05 19:01:34.514909] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:17.080 [2024-12-05 19:01:34.514954] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:17.341 "name": "raid_bdev1", 00:11:17.341 "uuid": "977c5b1f-d8aa-4a24-ba19-a43b28b8bc11", 00:11:17.341 "strip_size_kb": 0, 00:11:17.341 "state": "online", 00:11:17.341 "raid_level": "raid1", 00:11:17.341 "superblock": false, 00:11:17.341 "num_base_bdevs": 2, 00:11:17.341 "num_base_bdevs_discovered": 2, 00:11:17.341 "num_base_bdevs_operational": 2, 00:11:17.341 "base_bdevs_list": [ 00:11:17.341 { 00:11:17.341 "name": "spare", 00:11:17.341 "uuid": "717f5d56-410e-5dc7-8c1f-3e01f30852a9", 00:11:17.341 "is_configured": true, 00:11:17.341 "data_offset": 0, 00:11:17.341 "data_size": 65536 00:11:17.341 }, 00:11:17.341 { 00:11:17.341 "name": "BaseBdev2", 00:11:17.341 "uuid": "d32a3e77-b327-5474-a9a6-73d122557ebe", 00:11:17.341 "is_configured": true, 00:11:17.341 "data_offset": 0, 00:11:17.341 "data_size": 65536 00:11:17.341 } 00:11:17.341 ] 00:11:17.341 }' 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:17.341 "name": "raid_bdev1", 00:11:17.341 "uuid": "977c5b1f-d8aa-4a24-ba19-a43b28b8bc11", 00:11:17.341 "strip_size_kb": 0, 00:11:17.341 "state": "online", 00:11:17.341 "raid_level": "raid1", 00:11:17.341 "superblock": false, 00:11:17.341 "num_base_bdevs": 2, 00:11:17.341 "num_base_bdevs_discovered": 2, 00:11:17.341 "num_base_bdevs_operational": 2, 00:11:17.341 "base_bdevs_list": [ 00:11:17.341 { 00:11:17.341 "name": "spare", 00:11:17.341 "uuid": "717f5d56-410e-5dc7-8c1f-3e01f30852a9", 00:11:17.341 "is_configured": true, 00:11:17.341 "data_offset": 0, 00:11:17.341 "data_size": 65536 00:11:17.341 }, 00:11:17.341 { 00:11:17.341 "name": "BaseBdev2", 00:11:17.341 "uuid": "d32a3e77-b327-5474-a9a6-73d122557ebe", 00:11:17.341 "is_configured": true, 00:11:17.341 "data_offset": 0, 00:11:17.341 "data_size": 65536 00:11:17.341 } 00:11:17.341 ] 00:11:17.341 }' 00:11:17.341 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:17.601 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:17.601 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:17.601 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:17.601 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:17.602 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:17.602 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:17.602 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:17.602 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:17.602 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:17.602 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:17.602 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:17.602 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:17.602 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:17.602 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:17.602 19:01:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.602 19:01:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.602 19:01:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.602 19:01:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.602 19:01:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:17.602 "name": "raid_bdev1", 00:11:17.602 "uuid": "977c5b1f-d8aa-4a24-ba19-a43b28b8bc11", 00:11:17.602 "strip_size_kb": 0, 00:11:17.602 "state": "online", 00:11:17.602 "raid_level": "raid1", 00:11:17.602 "superblock": false, 00:11:17.602 "num_base_bdevs": 2, 00:11:17.602 "num_base_bdevs_discovered": 2, 00:11:17.602 "num_base_bdevs_operational": 2, 00:11:17.602 "base_bdevs_list": [ 00:11:17.602 { 00:11:17.602 "name": "spare", 00:11:17.602 "uuid": "717f5d56-410e-5dc7-8c1f-3e01f30852a9", 00:11:17.602 "is_configured": true, 00:11:17.602 "data_offset": 0, 00:11:17.602 "data_size": 65536 00:11:17.602 }, 00:11:17.602 { 00:11:17.602 "name": "BaseBdev2", 00:11:17.602 "uuid": "d32a3e77-b327-5474-a9a6-73d122557ebe", 00:11:17.602 "is_configured": true, 00:11:17.602 "data_offset": 0, 00:11:17.602 "data_size": 65536 00:11:17.602 } 00:11:17.602 ] 00:11:17.602 }' 00:11:17.602 19:01:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:17.602 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.861 19:01:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:17.861 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.861 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:17.861 [2024-12-05 19:01:35.401987] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:17.861 [2024-12-05 19:01:35.402061] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:17.861 [2024-12-05 19:01:35.402159] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:17.861 [2024-12-05 19:01:35.402254] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:17.861 [2024-12-05 19:01:35.402300] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:17.861 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:17.861 19:01:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.861 19:01:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:11:17.861 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:17.861 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:18.127 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:18.127 19:01:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:18.127 19:01:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:18.127 19:01:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:11:18.127 19:01:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:11:18.127 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:18.127 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:11:18.127 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:18.127 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:18.127 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:18.127 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:11:18.127 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:18.127 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:18.127 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:11:18.127 /dev/nbd0 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:18.386 1+0 records in 00:11:18.386 1+0 records out 00:11:18.386 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000294712 s, 13.9 MB/s 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:11:18.386 /dev/nbd1 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:18.386 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:18.646 1+0 records in 00:11:18.646 1+0 records out 00:11:18.646 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000433638 s, 9.4 MB/s 00:11:18.646 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:18.646 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:11:18.646 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:18.646 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:18.646 19:01:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:11:18.646 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:18.646 19:01:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:18.646 19:01:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:11:18.646 19:01:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:11:18.646 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:18.646 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:18.646 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:18.646 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:11:18.646 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:18.646 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 84766 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 84766 ']' 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 84766 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:18.906 19:01:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 84766 00:11:19.166 killing process with pid 84766 00:11:19.166 Received shutdown signal, test time was about 60.000000 seconds 00:11:19.166 00:11:19.166 Latency(us) 00:11:19.166 [2024-12-05T19:01:36.725Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:19.166 [2024-12-05T19:01:36.725Z] =================================================================================================================== 00:11:19.166 [2024-12-05T19:01:36.725Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:11:19.166 19:01:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:19.166 19:01:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:19.166 19:01:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 84766' 00:11:19.166 19:01:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@973 -- # kill 84766 00:11:19.166 [2024-12-05 19:01:36.476911] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:19.166 19:01:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@978 -- # wait 84766 00:11:19.166 [2024-12-05 19:01:36.507250] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:19.166 19:01:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:11:19.166 00:11:19.166 real 0m13.156s 00:11:19.166 user 0m15.420s 00:11:19.166 sys 0m2.617s 00:11:19.166 ************************************ 00:11:19.166 END TEST raid_rebuild_test 00:11:19.166 ************************************ 00:11:19.166 19:01:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:19.166 19:01:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.426 19:01:36 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 2 true false true 00:11:19.426 19:01:36 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:11:19.426 19:01:36 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:19.426 19:01:36 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:19.426 ************************************ 00:11:19.426 START TEST raid_rebuild_test_sb 00:11:19.426 ************************************ 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=85090 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 85090 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 85090 ']' 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:19.426 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:19.426 19:01:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.426 [2024-12-05 19:01:36.879569] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:11:19.426 [2024-12-05 19:01:36.879822] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85090 ] 00:11:19.426 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:19.426 Zero copy mechanism will not be used. 00:11:19.686 [2024-12-05 19:01:37.034418] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:19.686 [2024-12-05 19:01:37.059215] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:19.686 [2024-12-05 19:01:37.102488] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:19.686 [2024-12-05 19:01:37.102602] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.255 BaseBdev1_malloc 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.255 [2024-12-05 19:01:37.706692] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:20.255 [2024-12-05 19:01:37.706834] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:20.255 [2024-12-05 19:01:37.706874] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:20.255 [2024-12-05 19:01:37.706901] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:20.255 [2024-12-05 19:01:37.709048] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:20.255 [2024-12-05 19:01:37.709083] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:20.255 BaseBdev1 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.255 BaseBdev2_malloc 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.255 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.255 [2024-12-05 19:01:37.735453] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:20.255 [2024-12-05 19:01:37.735504] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:20.255 [2024-12-05 19:01:37.735527] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:20.256 [2024-12-05 19:01:37.735536] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:20.256 [2024-12-05 19:01:37.737678] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:20.256 [2024-12-05 19:01:37.737783] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:20.256 BaseBdev2 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.256 spare_malloc 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.256 spare_delay 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.256 [2024-12-05 19:01:37.776239] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:20.256 [2024-12-05 19:01:37.776284] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:20.256 [2024-12-05 19:01:37.776301] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:20.256 [2024-12-05 19:01:37.776309] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:20.256 [2024-12-05 19:01:37.778400] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:20.256 [2024-12-05 19:01:37.778432] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:20.256 spare 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.256 [2024-12-05 19:01:37.784268] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:20.256 [2024-12-05 19:01:37.786047] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:20.256 [2024-12-05 19:01:37.786222] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:20.256 [2024-12-05 19:01:37.786235] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:20.256 [2024-12-05 19:01:37.786517] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:11:20.256 [2024-12-05 19:01:37.786666] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:20.256 [2024-12-05 19:01:37.786679] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:20.256 [2024-12-05 19:01:37.786819] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.256 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.515 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.515 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:20.515 "name": "raid_bdev1", 00:11:20.515 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:20.515 "strip_size_kb": 0, 00:11:20.515 "state": "online", 00:11:20.515 "raid_level": "raid1", 00:11:20.515 "superblock": true, 00:11:20.515 "num_base_bdevs": 2, 00:11:20.515 "num_base_bdevs_discovered": 2, 00:11:20.515 "num_base_bdevs_operational": 2, 00:11:20.515 "base_bdevs_list": [ 00:11:20.515 { 00:11:20.515 "name": "BaseBdev1", 00:11:20.515 "uuid": "efe9c528-794d-5e27-9c73-24fc78344fd7", 00:11:20.515 "is_configured": true, 00:11:20.515 "data_offset": 2048, 00:11:20.515 "data_size": 63488 00:11:20.515 }, 00:11:20.515 { 00:11:20.515 "name": "BaseBdev2", 00:11:20.515 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:20.515 "is_configured": true, 00:11:20.515 "data_offset": 2048, 00:11:20.515 "data_size": 63488 00:11:20.515 } 00:11:20.515 ] 00:11:20.515 }' 00:11:20.515 19:01:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:20.515 19:01:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:20.775 [2024-12-05 19:01:38.227899] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:20.775 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:11:21.035 [2024-12-05 19:01:38.495265] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:11:21.035 /dev/nbd0 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:21.035 1+0 records in 00:11:21.035 1+0 records out 00:11:21.035 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000287643 s, 14.2 MB/s 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:11:21.035 19:01:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:11:25.232 63488+0 records in 00:11:25.232 63488+0 records out 00:11:25.232 32505856 bytes (33 MB, 31 MiB) copied, 3.83375 s, 8.5 MB/s 00:11:25.232 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:25.232 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:25.232 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:25.232 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:25.232 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:25.233 [2024-12-05 19:01:42.583499] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:25.233 [2024-12-05 19:01:42.615550] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:25.233 "name": "raid_bdev1", 00:11:25.233 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:25.233 "strip_size_kb": 0, 00:11:25.233 "state": "online", 00:11:25.233 "raid_level": "raid1", 00:11:25.233 "superblock": true, 00:11:25.233 "num_base_bdevs": 2, 00:11:25.233 "num_base_bdevs_discovered": 1, 00:11:25.233 "num_base_bdevs_operational": 1, 00:11:25.233 "base_bdevs_list": [ 00:11:25.233 { 00:11:25.233 "name": null, 00:11:25.233 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:25.233 "is_configured": false, 00:11:25.233 "data_offset": 0, 00:11:25.233 "data_size": 63488 00:11:25.233 }, 00:11:25.233 { 00:11:25.233 "name": "BaseBdev2", 00:11:25.233 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:25.233 "is_configured": true, 00:11:25.233 "data_offset": 2048, 00:11:25.233 "data_size": 63488 00:11:25.233 } 00:11:25.233 ] 00:11:25.233 }' 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:25.233 19:01:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:25.493 19:01:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:25.493 19:01:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:25.493 19:01:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:25.493 [2024-12-05 19:01:43.003007] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:25.493 [2024-12-05 19:01:43.008227] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e280 00:11:25.493 19:01:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:25.493 19:01:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:25.493 [2024-12-05 19:01:43.010153] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:26.875 "name": "raid_bdev1", 00:11:26.875 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:26.875 "strip_size_kb": 0, 00:11:26.875 "state": "online", 00:11:26.875 "raid_level": "raid1", 00:11:26.875 "superblock": true, 00:11:26.875 "num_base_bdevs": 2, 00:11:26.875 "num_base_bdevs_discovered": 2, 00:11:26.875 "num_base_bdevs_operational": 2, 00:11:26.875 "process": { 00:11:26.875 "type": "rebuild", 00:11:26.875 "target": "spare", 00:11:26.875 "progress": { 00:11:26.875 "blocks": 20480, 00:11:26.875 "percent": 32 00:11:26.875 } 00:11:26.875 }, 00:11:26.875 "base_bdevs_list": [ 00:11:26.875 { 00:11:26.875 "name": "spare", 00:11:26.875 "uuid": "5331160c-0793-5939-88c4-ceab1d96641a", 00:11:26.875 "is_configured": true, 00:11:26.875 "data_offset": 2048, 00:11:26.875 "data_size": 63488 00:11:26.875 }, 00:11:26.875 { 00:11:26.875 "name": "BaseBdev2", 00:11:26.875 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:26.875 "is_configured": true, 00:11:26.875 "data_offset": 2048, 00:11:26.875 "data_size": 63488 00:11:26.875 } 00:11:26.875 ] 00:11:26.875 }' 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:26.875 [2024-12-05 19:01:44.126964] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:26.875 [2024-12-05 19:01:44.214850] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:26.875 [2024-12-05 19:01:44.214981] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:26.875 [2024-12-05 19:01:44.215007] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:26.875 [2024-12-05 19:01:44.215018] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:26.875 "name": "raid_bdev1", 00:11:26.875 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:26.875 "strip_size_kb": 0, 00:11:26.875 "state": "online", 00:11:26.875 "raid_level": "raid1", 00:11:26.875 "superblock": true, 00:11:26.875 "num_base_bdevs": 2, 00:11:26.875 "num_base_bdevs_discovered": 1, 00:11:26.875 "num_base_bdevs_operational": 1, 00:11:26.875 "base_bdevs_list": [ 00:11:26.875 { 00:11:26.875 "name": null, 00:11:26.875 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:26.875 "is_configured": false, 00:11:26.875 "data_offset": 0, 00:11:26.875 "data_size": 63488 00:11:26.875 }, 00:11:26.875 { 00:11:26.875 "name": "BaseBdev2", 00:11:26.875 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:26.875 "is_configured": true, 00:11:26.875 "data_offset": 2048, 00:11:26.875 "data_size": 63488 00:11:26.875 } 00:11:26.875 ] 00:11:26.875 }' 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:26.875 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.135 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:27.135 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:27.135 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:27.135 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:27.135 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:27.135 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:27.135 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.135 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.135 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:27.135 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.135 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:27.135 "name": "raid_bdev1", 00:11:27.135 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:27.135 "strip_size_kb": 0, 00:11:27.135 "state": "online", 00:11:27.135 "raid_level": "raid1", 00:11:27.135 "superblock": true, 00:11:27.135 "num_base_bdevs": 2, 00:11:27.135 "num_base_bdevs_discovered": 1, 00:11:27.135 "num_base_bdevs_operational": 1, 00:11:27.135 "base_bdevs_list": [ 00:11:27.135 { 00:11:27.135 "name": null, 00:11:27.135 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:27.135 "is_configured": false, 00:11:27.135 "data_offset": 0, 00:11:27.135 "data_size": 63488 00:11:27.135 }, 00:11:27.135 { 00:11:27.135 "name": "BaseBdev2", 00:11:27.135 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:27.135 "is_configured": true, 00:11:27.135 "data_offset": 2048, 00:11:27.135 "data_size": 63488 00:11:27.135 } 00:11:27.135 ] 00:11:27.135 }' 00:11:27.135 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:27.394 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:27.394 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:27.394 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:27.394 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:27.394 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:27.394 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:27.394 [2024-12-05 19:01:44.751016] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:27.394 [2024-12-05 19:01:44.756033] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e350 00:11:27.394 19:01:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:27.394 19:01:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:27.394 [2024-12-05 19:01:44.757998] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:28.331 "name": "raid_bdev1", 00:11:28.331 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:28.331 "strip_size_kb": 0, 00:11:28.331 "state": "online", 00:11:28.331 "raid_level": "raid1", 00:11:28.331 "superblock": true, 00:11:28.331 "num_base_bdevs": 2, 00:11:28.331 "num_base_bdevs_discovered": 2, 00:11:28.331 "num_base_bdevs_operational": 2, 00:11:28.331 "process": { 00:11:28.331 "type": "rebuild", 00:11:28.331 "target": "spare", 00:11:28.331 "progress": { 00:11:28.331 "blocks": 20480, 00:11:28.331 "percent": 32 00:11:28.331 } 00:11:28.331 }, 00:11:28.331 "base_bdevs_list": [ 00:11:28.331 { 00:11:28.331 "name": "spare", 00:11:28.331 "uuid": "5331160c-0793-5939-88c4-ceab1d96641a", 00:11:28.331 "is_configured": true, 00:11:28.331 "data_offset": 2048, 00:11:28.331 "data_size": 63488 00:11:28.331 }, 00:11:28.331 { 00:11:28.331 "name": "BaseBdev2", 00:11:28.331 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:28.331 "is_configured": true, 00:11:28.331 "data_offset": 2048, 00:11:28.331 "data_size": 63488 00:11:28.331 } 00:11:28.331 ] 00:11:28.331 }' 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:11:28.331 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=295 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:28.331 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:28.590 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:28.590 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:28.590 19:01:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:28.590 19:01:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:28.590 19:01:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:28.590 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:28.590 "name": "raid_bdev1", 00:11:28.590 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:28.590 "strip_size_kb": 0, 00:11:28.590 "state": "online", 00:11:28.590 "raid_level": "raid1", 00:11:28.590 "superblock": true, 00:11:28.590 "num_base_bdevs": 2, 00:11:28.590 "num_base_bdevs_discovered": 2, 00:11:28.590 "num_base_bdevs_operational": 2, 00:11:28.590 "process": { 00:11:28.590 "type": "rebuild", 00:11:28.590 "target": "spare", 00:11:28.590 "progress": { 00:11:28.590 "blocks": 22528, 00:11:28.590 "percent": 35 00:11:28.590 } 00:11:28.590 }, 00:11:28.590 "base_bdevs_list": [ 00:11:28.590 { 00:11:28.590 "name": "spare", 00:11:28.590 "uuid": "5331160c-0793-5939-88c4-ceab1d96641a", 00:11:28.590 "is_configured": true, 00:11:28.590 "data_offset": 2048, 00:11:28.590 "data_size": 63488 00:11:28.590 }, 00:11:28.590 { 00:11:28.590 "name": "BaseBdev2", 00:11:28.590 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:28.590 "is_configured": true, 00:11:28.590 "data_offset": 2048, 00:11:28.591 "data_size": 63488 00:11:28.591 } 00:11:28.591 ] 00:11:28.591 }' 00:11:28.591 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:28.591 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:28.591 19:01:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:28.591 19:01:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:28.591 19:01:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:29.527 19:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:29.527 19:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:29.527 19:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:29.527 19:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:29.527 19:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:29.527 19:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:29.527 19:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:29.527 19:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:29.527 19:01:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:29.527 19:01:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:29.527 19:01:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:29.527 19:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:29.527 "name": "raid_bdev1", 00:11:29.527 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:29.527 "strip_size_kb": 0, 00:11:29.527 "state": "online", 00:11:29.527 "raid_level": "raid1", 00:11:29.527 "superblock": true, 00:11:29.527 "num_base_bdevs": 2, 00:11:29.527 "num_base_bdevs_discovered": 2, 00:11:29.527 "num_base_bdevs_operational": 2, 00:11:29.527 "process": { 00:11:29.527 "type": "rebuild", 00:11:29.527 "target": "spare", 00:11:29.527 "progress": { 00:11:29.527 "blocks": 45056, 00:11:29.527 "percent": 70 00:11:29.527 } 00:11:29.527 }, 00:11:29.527 "base_bdevs_list": [ 00:11:29.528 { 00:11:29.528 "name": "spare", 00:11:29.528 "uuid": "5331160c-0793-5939-88c4-ceab1d96641a", 00:11:29.528 "is_configured": true, 00:11:29.528 "data_offset": 2048, 00:11:29.528 "data_size": 63488 00:11:29.528 }, 00:11:29.528 { 00:11:29.528 "name": "BaseBdev2", 00:11:29.528 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:29.528 "is_configured": true, 00:11:29.528 "data_offset": 2048, 00:11:29.528 "data_size": 63488 00:11:29.528 } 00:11:29.528 ] 00:11:29.528 }' 00:11:29.528 19:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:29.787 19:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:29.787 19:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:29.787 19:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:29.787 19:01:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:30.356 [2024-12-05 19:01:47.868460] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:30.356 [2024-12-05 19:01:47.868629] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:30.356 [2024-12-05 19:01:47.868807] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:30.926 "name": "raid_bdev1", 00:11:30.926 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:30.926 "strip_size_kb": 0, 00:11:30.926 "state": "online", 00:11:30.926 "raid_level": "raid1", 00:11:30.926 "superblock": true, 00:11:30.926 "num_base_bdevs": 2, 00:11:30.926 "num_base_bdevs_discovered": 2, 00:11:30.926 "num_base_bdevs_operational": 2, 00:11:30.926 "base_bdevs_list": [ 00:11:30.926 { 00:11:30.926 "name": "spare", 00:11:30.926 "uuid": "5331160c-0793-5939-88c4-ceab1d96641a", 00:11:30.926 "is_configured": true, 00:11:30.926 "data_offset": 2048, 00:11:30.926 "data_size": 63488 00:11:30.926 }, 00:11:30.926 { 00:11:30.926 "name": "BaseBdev2", 00:11:30.926 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:30.926 "is_configured": true, 00:11:30.926 "data_offset": 2048, 00:11:30.926 "data_size": 63488 00:11:30.926 } 00:11:30.926 ] 00:11:30.926 }' 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:30.926 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:30.926 "name": "raid_bdev1", 00:11:30.927 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:30.927 "strip_size_kb": 0, 00:11:30.927 "state": "online", 00:11:30.927 "raid_level": "raid1", 00:11:30.927 "superblock": true, 00:11:30.927 "num_base_bdevs": 2, 00:11:30.927 "num_base_bdevs_discovered": 2, 00:11:30.927 "num_base_bdevs_operational": 2, 00:11:30.927 "base_bdevs_list": [ 00:11:30.927 { 00:11:30.927 "name": "spare", 00:11:30.927 "uuid": "5331160c-0793-5939-88c4-ceab1d96641a", 00:11:30.927 "is_configured": true, 00:11:30.927 "data_offset": 2048, 00:11:30.927 "data_size": 63488 00:11:30.927 }, 00:11:30.927 { 00:11:30.927 "name": "BaseBdev2", 00:11:30.927 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:30.927 "is_configured": true, 00:11:30.927 "data_offset": 2048, 00:11:30.927 "data_size": 63488 00:11:30.927 } 00:11:30.927 ] 00:11:30.927 }' 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:30.927 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.186 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:31.186 "name": "raid_bdev1", 00:11:31.186 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:31.186 "strip_size_kb": 0, 00:11:31.186 "state": "online", 00:11:31.186 "raid_level": "raid1", 00:11:31.186 "superblock": true, 00:11:31.186 "num_base_bdevs": 2, 00:11:31.186 "num_base_bdevs_discovered": 2, 00:11:31.186 "num_base_bdevs_operational": 2, 00:11:31.186 "base_bdevs_list": [ 00:11:31.186 { 00:11:31.186 "name": "spare", 00:11:31.186 "uuid": "5331160c-0793-5939-88c4-ceab1d96641a", 00:11:31.186 "is_configured": true, 00:11:31.186 "data_offset": 2048, 00:11:31.186 "data_size": 63488 00:11:31.186 }, 00:11:31.186 { 00:11:31.186 "name": "BaseBdev2", 00:11:31.186 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:31.186 "is_configured": true, 00:11:31.186 "data_offset": 2048, 00:11:31.186 "data_size": 63488 00:11:31.186 } 00:11:31.186 ] 00:11:31.186 }' 00:11:31.186 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:31.186 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.446 [2024-12-05 19:01:48.875712] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:31.446 [2024-12-05 19:01:48.875796] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:31.446 [2024-12-05 19:01:48.875913] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:31.446 [2024-12-05 19:01:48.876014] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:31.446 [2024-12-05 19:01:48.876084] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:31.446 19:01:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:11:31.705 /dev/nbd0 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:31.706 1+0 records in 00:11:31.706 1+0 records out 00:11:31.706 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000623023 s, 6.6 MB/s 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:31.706 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:11:31.965 /dev/nbd1 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:31.965 1+0 records in 00:11:31.965 1+0 records out 00:11:31.965 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000426468 s, 9.6 MB/s 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:11:31.965 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:31.966 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:31.966 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:31.966 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:11:31.966 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:31.966 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:32.226 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:32.226 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:32.226 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:32.226 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:32.226 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:32.226 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:32.226 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:32.226 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:32.226 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:32.226 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.486 [2024-12-05 19:01:49.914022] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:32.486 [2024-12-05 19:01:49.914090] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:32.486 [2024-12-05 19:01:49.914113] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:32.486 [2024-12-05 19:01:49.914128] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:32.486 [2024-12-05 19:01:49.916278] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:32.486 [2024-12-05 19:01:49.916337] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:32.486 [2024-12-05 19:01:49.916439] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:32.486 [2024-12-05 19:01:49.916489] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:32.486 [2024-12-05 19:01:49.916622] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:32.486 spare 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.486 19:01:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.486 [2024-12-05 19:01:50.016550] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:11:32.486 [2024-12-05 19:01:50.016631] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:32.486 [2024-12-05 19:01:50.016980] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cae960 00:11:32.486 [2024-12-05 19:01:50.017152] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:11:32.486 [2024-12-05 19:01:50.017168] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:11:32.486 [2024-12-05 19:01:50.017308] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:32.486 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.486 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:32.486 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:32.486 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:32.486 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:32.487 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:32.487 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:32.487 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:32.487 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:32.487 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:32.487 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:32.487 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:32.487 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:32.487 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:32.487 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:32.487 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:32.747 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:32.747 "name": "raid_bdev1", 00:11:32.747 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:32.747 "strip_size_kb": 0, 00:11:32.747 "state": "online", 00:11:32.747 "raid_level": "raid1", 00:11:32.747 "superblock": true, 00:11:32.747 "num_base_bdevs": 2, 00:11:32.747 "num_base_bdevs_discovered": 2, 00:11:32.747 "num_base_bdevs_operational": 2, 00:11:32.747 "base_bdevs_list": [ 00:11:32.747 { 00:11:32.747 "name": "spare", 00:11:32.747 "uuid": "5331160c-0793-5939-88c4-ceab1d96641a", 00:11:32.747 "is_configured": true, 00:11:32.747 "data_offset": 2048, 00:11:32.747 "data_size": 63488 00:11:32.747 }, 00:11:32.747 { 00:11:32.747 "name": "BaseBdev2", 00:11:32.747 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:32.747 "is_configured": true, 00:11:32.747 "data_offset": 2048, 00:11:32.747 "data_size": 63488 00:11:32.747 } 00:11:32.747 ] 00:11:32.747 }' 00:11:32.747 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:32.747 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.006 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:33.006 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:33.006 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:33.006 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:33.006 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:33.006 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.006 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:33.006 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.006 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.006 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.006 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:33.006 "name": "raid_bdev1", 00:11:33.006 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:33.006 "strip_size_kb": 0, 00:11:33.006 "state": "online", 00:11:33.006 "raid_level": "raid1", 00:11:33.006 "superblock": true, 00:11:33.006 "num_base_bdevs": 2, 00:11:33.006 "num_base_bdevs_discovered": 2, 00:11:33.006 "num_base_bdevs_operational": 2, 00:11:33.006 "base_bdevs_list": [ 00:11:33.006 { 00:11:33.006 "name": "spare", 00:11:33.006 "uuid": "5331160c-0793-5939-88c4-ceab1d96641a", 00:11:33.006 "is_configured": true, 00:11:33.006 "data_offset": 2048, 00:11:33.006 "data_size": 63488 00:11:33.006 }, 00:11:33.006 { 00:11:33.006 "name": "BaseBdev2", 00:11:33.006 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:33.006 "is_configured": true, 00:11:33.006 "data_offset": 2048, 00:11:33.006 "data_size": 63488 00:11:33.006 } 00:11:33.006 ] 00:11:33.006 }' 00:11:33.006 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:33.006 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:33.006 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.266 [2024-12-05 19:01:50.628810] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:33.266 "name": "raid_bdev1", 00:11:33.266 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:33.266 "strip_size_kb": 0, 00:11:33.266 "state": "online", 00:11:33.266 "raid_level": "raid1", 00:11:33.266 "superblock": true, 00:11:33.266 "num_base_bdevs": 2, 00:11:33.266 "num_base_bdevs_discovered": 1, 00:11:33.266 "num_base_bdevs_operational": 1, 00:11:33.266 "base_bdevs_list": [ 00:11:33.266 { 00:11:33.266 "name": null, 00:11:33.266 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:33.266 "is_configured": false, 00:11:33.266 "data_offset": 0, 00:11:33.266 "data_size": 63488 00:11:33.266 }, 00:11:33.266 { 00:11:33.266 "name": "BaseBdev2", 00:11:33.266 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:33.266 "is_configured": true, 00:11:33.266 "data_offset": 2048, 00:11:33.266 "data_size": 63488 00:11:33.266 } 00:11:33.266 ] 00:11:33.266 }' 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:33.266 19:01:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.525 19:01:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:33.526 19:01:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:33.526 19:01:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:33.526 [2024-12-05 19:01:51.048180] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:33.526 [2024-12-05 19:01:51.048373] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:33.526 [2024-12-05 19:01:51.048388] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:33.526 [2024-12-05 19:01:51.048444] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:33.526 [2024-12-05 19:01:51.053424] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caea30 00:11:33.526 19:01:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:33.526 19:01:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:11:33.526 [2024-12-05 19:01:51.055318] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:34.904 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:34.904 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:34.904 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:34.904 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:34.904 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:34.904 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.904 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:34.904 19:01:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.904 19:01:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.904 19:01:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.904 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:34.904 "name": "raid_bdev1", 00:11:34.904 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:34.904 "strip_size_kb": 0, 00:11:34.904 "state": "online", 00:11:34.904 "raid_level": "raid1", 00:11:34.904 "superblock": true, 00:11:34.904 "num_base_bdevs": 2, 00:11:34.904 "num_base_bdevs_discovered": 2, 00:11:34.904 "num_base_bdevs_operational": 2, 00:11:34.904 "process": { 00:11:34.904 "type": "rebuild", 00:11:34.905 "target": "spare", 00:11:34.905 "progress": { 00:11:34.905 "blocks": 20480, 00:11:34.905 "percent": 32 00:11:34.905 } 00:11:34.905 }, 00:11:34.905 "base_bdevs_list": [ 00:11:34.905 { 00:11:34.905 "name": "spare", 00:11:34.905 "uuid": "5331160c-0793-5939-88c4-ceab1d96641a", 00:11:34.905 "is_configured": true, 00:11:34.905 "data_offset": 2048, 00:11:34.905 "data_size": 63488 00:11:34.905 }, 00:11:34.905 { 00:11:34.905 "name": "BaseBdev2", 00:11:34.905 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:34.905 "is_configured": true, 00:11:34.905 "data_offset": 2048, 00:11:34.905 "data_size": 63488 00:11:34.905 } 00:11:34.905 ] 00:11:34.905 }' 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.905 [2024-12-05 19:01:52.215503] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:34.905 [2024-12-05 19:01:52.259352] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:34.905 [2024-12-05 19:01:52.259469] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:34.905 [2024-12-05 19:01:52.259514] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:34.905 [2024-12-05 19:01:52.259540] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:34.905 "name": "raid_bdev1", 00:11:34.905 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:34.905 "strip_size_kb": 0, 00:11:34.905 "state": "online", 00:11:34.905 "raid_level": "raid1", 00:11:34.905 "superblock": true, 00:11:34.905 "num_base_bdevs": 2, 00:11:34.905 "num_base_bdevs_discovered": 1, 00:11:34.905 "num_base_bdevs_operational": 1, 00:11:34.905 "base_bdevs_list": [ 00:11:34.905 { 00:11:34.905 "name": null, 00:11:34.905 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:34.905 "is_configured": false, 00:11:34.905 "data_offset": 0, 00:11:34.905 "data_size": 63488 00:11:34.905 }, 00:11:34.905 { 00:11:34.905 "name": "BaseBdev2", 00:11:34.905 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:34.905 "is_configured": true, 00:11:34.905 "data_offset": 2048, 00:11:34.905 "data_size": 63488 00:11:34.905 } 00:11:34.905 ] 00:11:34.905 }' 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:34.905 19:01:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.164 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:35.164 19:01:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:35.164 19:01:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:35.164 [2024-12-05 19:01:52.683550] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:35.164 [2024-12-05 19:01:52.683678] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:35.164 [2024-12-05 19:01:52.683726] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:11:35.164 [2024-12-05 19:01:52.683762] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:35.164 [2024-12-05 19:01:52.684254] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:35.164 [2024-12-05 19:01:52.684323] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:35.164 [2024-12-05 19:01:52.684452] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:35.164 [2024-12-05 19:01:52.684496] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:35.164 [2024-12-05 19:01:52.684562] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:35.164 [2024-12-05 19:01:52.684635] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:35.164 [2024-12-05 19:01:52.689100] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeb00 00:11:35.164 spare 00:11:35.165 19:01:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:35.165 19:01:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:11:35.165 [2024-12-05 19:01:52.691026] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:36.544 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:36.544 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:36.544 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:36.544 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:36.544 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:36.544 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:36.544 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:36.544 19:01:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:36.544 19:01:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:36.544 19:01:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:36.544 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:36.544 "name": "raid_bdev1", 00:11:36.544 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:36.544 "strip_size_kb": 0, 00:11:36.545 "state": "online", 00:11:36.545 "raid_level": "raid1", 00:11:36.545 "superblock": true, 00:11:36.545 "num_base_bdevs": 2, 00:11:36.545 "num_base_bdevs_discovered": 2, 00:11:36.545 "num_base_bdevs_operational": 2, 00:11:36.545 "process": { 00:11:36.545 "type": "rebuild", 00:11:36.545 "target": "spare", 00:11:36.545 "progress": { 00:11:36.545 "blocks": 20480, 00:11:36.545 "percent": 32 00:11:36.545 } 00:11:36.545 }, 00:11:36.545 "base_bdevs_list": [ 00:11:36.545 { 00:11:36.545 "name": "spare", 00:11:36.545 "uuid": "5331160c-0793-5939-88c4-ceab1d96641a", 00:11:36.545 "is_configured": true, 00:11:36.545 "data_offset": 2048, 00:11:36.545 "data_size": 63488 00:11:36.545 }, 00:11:36.545 { 00:11:36.545 "name": "BaseBdev2", 00:11:36.545 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:36.545 "is_configured": true, 00:11:36.545 "data_offset": 2048, 00:11:36.545 "data_size": 63488 00:11:36.545 } 00:11:36.545 ] 00:11:36.545 }' 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:36.545 [2024-12-05 19:01:53.847293] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:36.545 [2024-12-05 19:01:53.895135] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:36.545 [2024-12-05 19:01:53.895202] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:36.545 [2024-12-05 19:01:53.895219] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:36.545 [2024-12-05 19:01:53.895230] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:36.545 "name": "raid_bdev1", 00:11:36.545 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:36.545 "strip_size_kb": 0, 00:11:36.545 "state": "online", 00:11:36.545 "raid_level": "raid1", 00:11:36.545 "superblock": true, 00:11:36.545 "num_base_bdevs": 2, 00:11:36.545 "num_base_bdevs_discovered": 1, 00:11:36.545 "num_base_bdevs_operational": 1, 00:11:36.545 "base_bdevs_list": [ 00:11:36.545 { 00:11:36.545 "name": null, 00:11:36.545 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:36.545 "is_configured": false, 00:11:36.545 "data_offset": 0, 00:11:36.545 "data_size": 63488 00:11:36.545 }, 00:11:36.545 { 00:11:36.545 "name": "BaseBdev2", 00:11:36.545 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:36.545 "is_configured": true, 00:11:36.545 "data_offset": 2048, 00:11:36.545 "data_size": 63488 00:11:36.545 } 00:11:36.545 ] 00:11:36.545 }' 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:36.545 19:01:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:36.805 19:01:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:36.805 19:01:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:36.805 19:01:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:36.805 19:01:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:36.805 19:01:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:36.805 19:01:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:36.805 19:01:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:36.805 19:01:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:36.805 19:01:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:36.805 19:01:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.067 19:01:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:37.067 "name": "raid_bdev1", 00:11:37.067 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:37.067 "strip_size_kb": 0, 00:11:37.067 "state": "online", 00:11:37.067 "raid_level": "raid1", 00:11:37.067 "superblock": true, 00:11:37.067 "num_base_bdevs": 2, 00:11:37.067 "num_base_bdevs_discovered": 1, 00:11:37.067 "num_base_bdevs_operational": 1, 00:11:37.067 "base_bdevs_list": [ 00:11:37.067 { 00:11:37.067 "name": null, 00:11:37.067 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:37.067 "is_configured": false, 00:11:37.067 "data_offset": 0, 00:11:37.067 "data_size": 63488 00:11:37.067 }, 00:11:37.067 { 00:11:37.067 "name": "BaseBdev2", 00:11:37.067 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:37.067 "is_configured": true, 00:11:37.067 "data_offset": 2048, 00:11:37.067 "data_size": 63488 00:11:37.067 } 00:11:37.067 ] 00:11:37.067 }' 00:11:37.067 19:01:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:37.067 19:01:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:37.067 19:01:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:37.067 19:01:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:37.067 19:01:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:11:37.067 19:01:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.067 19:01:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:37.067 19:01:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.067 19:01:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:37.067 19:01:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:37.067 19:01:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:37.067 [2024-12-05 19:01:54.454942] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:37.067 [2024-12-05 19:01:54.455055] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:37.067 [2024-12-05 19:01:54.455082] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:11:37.067 [2024-12-05 19:01:54.455096] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:37.067 [2024-12-05 19:01:54.455529] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:37.067 [2024-12-05 19:01:54.455551] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:37.067 [2024-12-05 19:01:54.455631] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:11:37.067 [2024-12-05 19:01:54.455652] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:37.067 [2024-12-05 19:01:54.455660] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:37.067 [2024-12-05 19:01:54.455703] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:11:37.067 BaseBdev1 00:11:37.067 19:01:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:37.067 19:01:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:38.026 "name": "raid_bdev1", 00:11:38.026 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:38.026 "strip_size_kb": 0, 00:11:38.026 "state": "online", 00:11:38.026 "raid_level": "raid1", 00:11:38.026 "superblock": true, 00:11:38.026 "num_base_bdevs": 2, 00:11:38.026 "num_base_bdevs_discovered": 1, 00:11:38.026 "num_base_bdevs_operational": 1, 00:11:38.026 "base_bdevs_list": [ 00:11:38.026 { 00:11:38.026 "name": null, 00:11:38.026 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:38.026 "is_configured": false, 00:11:38.026 "data_offset": 0, 00:11:38.026 "data_size": 63488 00:11:38.026 }, 00:11:38.026 { 00:11:38.026 "name": "BaseBdev2", 00:11:38.026 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:38.026 "is_configured": true, 00:11:38.026 "data_offset": 2048, 00:11:38.026 "data_size": 63488 00:11:38.026 } 00:11:38.026 ] 00:11:38.026 }' 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:38.026 19:01:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:38.594 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:38.594 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:38.594 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:38.594 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:38.594 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:38.594 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.594 19:01:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.594 19:01:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:38.594 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:38.594 19:01:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:38.594 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:38.594 "name": "raid_bdev1", 00:11:38.594 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:38.594 "strip_size_kb": 0, 00:11:38.594 "state": "online", 00:11:38.594 "raid_level": "raid1", 00:11:38.594 "superblock": true, 00:11:38.594 "num_base_bdevs": 2, 00:11:38.594 "num_base_bdevs_discovered": 1, 00:11:38.594 "num_base_bdevs_operational": 1, 00:11:38.594 "base_bdevs_list": [ 00:11:38.594 { 00:11:38.594 "name": null, 00:11:38.594 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:38.594 "is_configured": false, 00:11:38.594 "data_offset": 0, 00:11:38.594 "data_size": 63488 00:11:38.594 }, 00:11:38.594 { 00:11:38.594 "name": "BaseBdev2", 00:11:38.594 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:38.594 "is_configured": true, 00:11:38.594 "data_offset": 2048, 00:11:38.594 "data_size": 63488 00:11:38.594 } 00:11:38.594 ] 00:11:38.594 }' 00:11:38.594 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:38.594 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:38.594 19:01:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:38.594 [2024-12-05 19:01:56.021315] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:38.594 [2024-12-05 19:01:56.021531] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:38.594 [2024-12-05 19:01:56.021548] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:38.594 request: 00:11:38.594 { 00:11:38.594 "base_bdev": "BaseBdev1", 00:11:38.594 "raid_bdev": "raid_bdev1", 00:11:38.594 "method": "bdev_raid_add_base_bdev", 00:11:38.594 "req_id": 1 00:11:38.594 } 00:11:38.594 Got JSON-RPC error response 00:11:38.594 response: 00:11:38.594 { 00:11:38.594 "code": -22, 00:11:38.594 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:11:38.594 } 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:11:38.594 19:01:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:39.532 "name": "raid_bdev1", 00:11:39.532 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:39.532 "strip_size_kb": 0, 00:11:39.532 "state": "online", 00:11:39.532 "raid_level": "raid1", 00:11:39.532 "superblock": true, 00:11:39.532 "num_base_bdevs": 2, 00:11:39.532 "num_base_bdevs_discovered": 1, 00:11:39.532 "num_base_bdevs_operational": 1, 00:11:39.532 "base_bdevs_list": [ 00:11:39.532 { 00:11:39.532 "name": null, 00:11:39.532 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:39.532 "is_configured": false, 00:11:39.532 "data_offset": 0, 00:11:39.532 "data_size": 63488 00:11:39.532 }, 00:11:39.532 { 00:11:39.532 "name": "BaseBdev2", 00:11:39.532 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:39.532 "is_configured": true, 00:11:39.532 "data_offset": 2048, 00:11:39.532 "data_size": 63488 00:11:39.532 } 00:11:39.532 ] 00:11:39.532 }' 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:39.532 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.101 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:40.101 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:40.101 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:40.101 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:40.101 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:40.101 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:40.101 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:40.101 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:40.101 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.101 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:40.101 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:40.101 "name": "raid_bdev1", 00:11:40.101 "uuid": "0c96d17c-8574-4473-831f-e19db604843a", 00:11:40.101 "strip_size_kb": 0, 00:11:40.101 "state": "online", 00:11:40.101 "raid_level": "raid1", 00:11:40.101 "superblock": true, 00:11:40.101 "num_base_bdevs": 2, 00:11:40.101 "num_base_bdevs_discovered": 1, 00:11:40.102 "num_base_bdevs_operational": 1, 00:11:40.102 "base_bdevs_list": [ 00:11:40.102 { 00:11:40.102 "name": null, 00:11:40.102 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:40.102 "is_configured": false, 00:11:40.102 "data_offset": 0, 00:11:40.102 "data_size": 63488 00:11:40.102 }, 00:11:40.102 { 00:11:40.102 "name": "BaseBdev2", 00:11:40.102 "uuid": "002ffbed-9d0d-587e-9b68-ffb334bf48b3", 00:11:40.102 "is_configured": true, 00:11:40.102 "data_offset": 2048, 00:11:40.102 "data_size": 63488 00:11:40.102 } 00:11:40.102 ] 00:11:40.102 }' 00:11:40.102 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:40.102 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:40.102 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:40.102 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:40.102 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 85090 00:11:40.102 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 85090 ']' 00:11:40.102 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 85090 00:11:40.102 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:11:40.102 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:40.102 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85090 00:11:40.102 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:40.102 killing process with pid 85090 00:11:40.102 Received shutdown signal, test time was about 60.000000 seconds 00:11:40.102 00:11:40.102 Latency(us) 00:11:40.102 [2024-12-05T19:01:57.661Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:40.102 [2024-12-05T19:01:57.661Z] =================================================================================================================== 00:11:40.102 [2024-12-05T19:01:57.661Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:11:40.102 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:40.102 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85090' 00:11:40.102 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 85090 00:11:40.102 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 85090 00:11:40.102 [2024-12-05 19:01:57.587617] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:40.102 [2024-12-05 19:01:57.587763] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:40.102 [2024-12-05 19:01:57.587827] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:40.102 [2024-12-05 19:01:57.587837] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:11:40.102 [2024-12-05 19:01:57.618818] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:11:40.363 00:11:40.363 real 0m21.039s 00:11:40.363 user 0m25.741s 00:11:40.363 sys 0m3.476s 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:40.363 ************************************ 00:11:40.363 END TEST raid_rebuild_test_sb 00:11:40.363 ************************************ 00:11:40.363 19:01:57 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 2 false true true 00:11:40.363 19:01:57 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:11:40.363 19:01:57 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:40.363 19:01:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:40.363 ************************************ 00:11:40.363 START TEST raid_rebuild_test_io 00:11:40.363 ************************************ 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 false true true 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=85673 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 85673 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # '[' -z 85673 ']' 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:40.363 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:40.363 19:01:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:40.623 [2024-12-05 19:01:57.994913] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:11:40.623 [2024-12-05 19:01:57.995491] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85673 ] 00:11:40.623 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:40.623 Zero copy mechanism will not be used. 00:11:40.623 [2024-12-05 19:01:58.149921] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:40.623 [2024-12-05 19:01:58.174994] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:40.884 [2024-12-05 19:01:58.218601] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:40.884 [2024-12-05 19:01:58.218749] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # return 0 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.454 BaseBdev1_malloc 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.454 [2024-12-05 19:01:58.823404] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:41.454 [2024-12-05 19:01:58.823530] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:41.454 [2024-12-05 19:01:58.823583] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:41.454 [2024-12-05 19:01:58.823638] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:41.454 [2024-12-05 19:01:58.825827] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:41.454 [2024-12-05 19:01:58.825922] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:41.454 BaseBdev1 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.454 BaseBdev2_malloc 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.454 [2024-12-05 19:01:58.852163] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:41.454 [2024-12-05 19:01:58.852224] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:41.454 [2024-12-05 19:01:58.852251] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:41.454 [2024-12-05 19:01:58.852261] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:41.454 [2024-12-05 19:01:58.854408] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:41.454 [2024-12-05 19:01:58.854454] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:41.454 BaseBdev2 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.454 spare_malloc 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.454 spare_delay 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.454 [2024-12-05 19:01:58.892959] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:41.454 [2024-12-05 19:01:58.893012] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:41.454 [2024-12-05 19:01:58.893033] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:41.454 [2024-12-05 19:01:58.893043] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:41.454 [2024-12-05 19:01:58.895175] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:41.454 [2024-12-05 19:01:58.895227] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:41.454 spare 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.454 [2024-12-05 19:01:58.904985] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:41.454 [2024-12-05 19:01:58.906864] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:41.454 [2024-12-05 19:01:58.906960] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:41.454 [2024-12-05 19:01:58.906978] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:41.454 [2024-12-05 19:01:58.907258] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:11:41.454 [2024-12-05 19:01:58.907399] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:41.454 [2024-12-05 19:01:58.907419] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:41.454 [2024-12-05 19:01:58.907535] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:41.454 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:41.455 "name": "raid_bdev1", 00:11:41.455 "uuid": "65beaccb-3c7b-48a9-9d0c-f2423806130b", 00:11:41.455 "strip_size_kb": 0, 00:11:41.455 "state": "online", 00:11:41.455 "raid_level": "raid1", 00:11:41.455 "superblock": false, 00:11:41.455 "num_base_bdevs": 2, 00:11:41.455 "num_base_bdevs_discovered": 2, 00:11:41.455 "num_base_bdevs_operational": 2, 00:11:41.455 "base_bdevs_list": [ 00:11:41.455 { 00:11:41.455 "name": "BaseBdev1", 00:11:41.455 "uuid": "e6aca104-ec78-5468-8b1f-8a850a44e3f5", 00:11:41.455 "is_configured": true, 00:11:41.455 "data_offset": 0, 00:11:41.455 "data_size": 65536 00:11:41.455 }, 00:11:41.455 { 00:11:41.455 "name": "BaseBdev2", 00:11:41.455 "uuid": "b417b844-b703-58ff-98ba-17d31eb83e96", 00:11:41.455 "is_configured": true, 00:11:41.455 "data_offset": 0, 00:11:41.455 "data_size": 65536 00:11:41.455 } 00:11:41.455 ] 00:11:41.455 }' 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:41.455 19:01:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.025 [2024-12-05 19:01:59.364445] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.025 [2024-12-05 19:01:59.436084] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:42.025 "name": "raid_bdev1", 00:11:42.025 "uuid": "65beaccb-3c7b-48a9-9d0c-f2423806130b", 00:11:42.025 "strip_size_kb": 0, 00:11:42.025 "state": "online", 00:11:42.025 "raid_level": "raid1", 00:11:42.025 "superblock": false, 00:11:42.025 "num_base_bdevs": 2, 00:11:42.025 "num_base_bdevs_discovered": 1, 00:11:42.025 "num_base_bdevs_operational": 1, 00:11:42.025 "base_bdevs_list": [ 00:11:42.025 { 00:11:42.025 "name": null, 00:11:42.025 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:42.025 "is_configured": false, 00:11:42.025 "data_offset": 0, 00:11:42.025 "data_size": 65536 00:11:42.025 }, 00:11:42.025 { 00:11:42.025 "name": "BaseBdev2", 00:11:42.025 "uuid": "b417b844-b703-58ff-98ba-17d31eb83e96", 00:11:42.025 "is_configured": true, 00:11:42.025 "data_offset": 0, 00:11:42.025 "data_size": 65536 00:11:42.025 } 00:11:42.025 ] 00:11:42.025 }' 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:42.025 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.025 [2024-12-05 19:01:59.537064] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:11:42.025 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:42.025 Zero copy mechanism will not be used. 00:11:42.025 Running I/O for 60 seconds... 00:11:42.595 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:42.595 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:42.595 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.595 [2024-12-05 19:01:59.879795] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:42.595 19:01:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:42.595 19:01:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:42.595 [2024-12-05 19:01:59.931761] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:11:42.595 [2024-12-05 19:01:59.933759] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:42.595 [2024-12-05 19:02:00.046419] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:42.596 [2024-12-05 19:02:00.046898] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:42.855 [2024-12-05 19:02:00.271064] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:42.855 [2024-12-05 19:02:00.271348] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:43.114 [2024-12-05 19:02:00.493617] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:43.114 [2024-12-05 19:02:00.494020] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:43.374 212.00 IOPS, 636.00 MiB/s [2024-12-05T19:02:00.933Z] [2024-12-05 19:02:00.712741] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:43.374 [2024-12-05 19:02:00.713121] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:43.374 19:02:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:43.374 19:02:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:43.374 19:02:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:43.374 19:02:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:43.374 19:02:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:43.374 19:02:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:43.374 19:02:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:43.374 19:02:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:43.374 19:02:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:43.634 19:02:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:43.634 19:02:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:43.634 "name": "raid_bdev1", 00:11:43.634 "uuid": "65beaccb-3c7b-48a9-9d0c-f2423806130b", 00:11:43.634 "strip_size_kb": 0, 00:11:43.634 "state": "online", 00:11:43.634 "raid_level": "raid1", 00:11:43.634 "superblock": false, 00:11:43.634 "num_base_bdevs": 2, 00:11:43.634 "num_base_bdevs_discovered": 2, 00:11:43.634 "num_base_bdevs_operational": 2, 00:11:43.634 "process": { 00:11:43.634 "type": "rebuild", 00:11:43.634 "target": "spare", 00:11:43.634 "progress": { 00:11:43.634 "blocks": 10240, 00:11:43.634 "percent": 15 00:11:43.634 } 00:11:43.634 }, 00:11:43.634 "base_bdevs_list": [ 00:11:43.634 { 00:11:43.634 "name": "spare", 00:11:43.634 "uuid": "7abb63dc-21f5-5627-847b-265974781628", 00:11:43.634 "is_configured": true, 00:11:43.634 "data_offset": 0, 00:11:43.634 "data_size": 65536 00:11:43.634 }, 00:11:43.634 { 00:11:43.634 "name": "BaseBdev2", 00:11:43.634 "uuid": "b417b844-b703-58ff-98ba-17d31eb83e96", 00:11:43.634 "is_configured": true, 00:11:43.634 "data_offset": 0, 00:11:43.634 "data_size": 65536 00:11:43.634 } 00:11:43.634 ] 00:11:43.634 }' 00:11:43.634 19:02:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:43.634 19:02:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:43.634 19:02:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:43.634 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:43.634 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:43.634 19:02:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:43.634 19:02:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:43.634 [2024-12-05 19:02:01.051830] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:43.634 [2024-12-05 19:02:01.059078] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:43.634 [2024-12-05 19:02:01.159895] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:43.634 [2024-12-05 19:02:01.167149] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:43.634 [2024-12-05 19:02:01.167253] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:43.634 [2024-12-05 19:02:01.167274] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:43.634 [2024-12-05 19:02:01.178682] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000026d0 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:43.894 "name": "raid_bdev1", 00:11:43.894 "uuid": "65beaccb-3c7b-48a9-9d0c-f2423806130b", 00:11:43.894 "strip_size_kb": 0, 00:11:43.894 "state": "online", 00:11:43.894 "raid_level": "raid1", 00:11:43.894 "superblock": false, 00:11:43.894 "num_base_bdevs": 2, 00:11:43.894 "num_base_bdevs_discovered": 1, 00:11:43.894 "num_base_bdevs_operational": 1, 00:11:43.894 "base_bdevs_list": [ 00:11:43.894 { 00:11:43.894 "name": null, 00:11:43.894 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:43.894 "is_configured": false, 00:11:43.894 "data_offset": 0, 00:11:43.894 "data_size": 65536 00:11:43.894 }, 00:11:43.894 { 00:11:43.894 "name": "BaseBdev2", 00:11:43.894 "uuid": "b417b844-b703-58ff-98ba-17d31eb83e96", 00:11:43.894 "is_configured": true, 00:11:43.894 "data_offset": 0, 00:11:43.894 "data_size": 65536 00:11:43.894 } 00:11:43.894 ] 00:11:43.894 }' 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:43.894 19:02:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:44.154 182.50 IOPS, 547.50 MiB/s [2024-12-05T19:02:01.713Z] 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:44.154 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:44.154 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:44.154 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:44.154 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:44.154 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:44.154 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:44.154 19:02:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:44.154 19:02:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:44.154 19:02:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:44.154 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:44.154 "name": "raid_bdev1", 00:11:44.154 "uuid": "65beaccb-3c7b-48a9-9d0c-f2423806130b", 00:11:44.154 "strip_size_kb": 0, 00:11:44.154 "state": "online", 00:11:44.154 "raid_level": "raid1", 00:11:44.154 "superblock": false, 00:11:44.154 "num_base_bdevs": 2, 00:11:44.154 "num_base_bdevs_discovered": 1, 00:11:44.154 "num_base_bdevs_operational": 1, 00:11:44.154 "base_bdevs_list": [ 00:11:44.154 { 00:11:44.154 "name": null, 00:11:44.154 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:44.154 "is_configured": false, 00:11:44.154 "data_offset": 0, 00:11:44.154 "data_size": 65536 00:11:44.154 }, 00:11:44.154 { 00:11:44.154 "name": "BaseBdev2", 00:11:44.154 "uuid": "b417b844-b703-58ff-98ba-17d31eb83e96", 00:11:44.154 "is_configured": true, 00:11:44.154 "data_offset": 0, 00:11:44.154 "data_size": 65536 00:11:44.154 } 00:11:44.154 ] 00:11:44.154 }' 00:11:44.154 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:44.154 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:44.154 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:44.413 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:44.413 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:44.413 19:02:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:44.413 19:02:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:44.413 [2024-12-05 19:02:01.734027] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:44.413 19:02:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:44.413 19:02:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:44.413 [2024-12-05 19:02:01.777752] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:11:44.413 [2024-12-05 19:02:01.779681] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:44.413 [2024-12-05 19:02:01.896174] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:44.413 [2024-12-05 19:02:01.896621] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:44.674 [2024-12-05 19:02:02.108806] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:44.674 [2024-12-05 19:02:02.109086] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:44.934 [2024-12-05 19:02:02.422636] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:45.194 161.00 IOPS, 483.00 MiB/s [2024-12-05T19:02:02.753Z] [2024-12-05 19:02:02.628989] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:45.194 [2024-12-05 19:02:02.629171] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:45.455 "name": "raid_bdev1", 00:11:45.455 "uuid": "65beaccb-3c7b-48a9-9d0c-f2423806130b", 00:11:45.455 "strip_size_kb": 0, 00:11:45.455 "state": "online", 00:11:45.455 "raid_level": "raid1", 00:11:45.455 "superblock": false, 00:11:45.455 "num_base_bdevs": 2, 00:11:45.455 "num_base_bdevs_discovered": 2, 00:11:45.455 "num_base_bdevs_operational": 2, 00:11:45.455 "process": { 00:11:45.455 "type": "rebuild", 00:11:45.455 "target": "spare", 00:11:45.455 "progress": { 00:11:45.455 "blocks": 10240, 00:11:45.455 "percent": 15 00:11:45.455 } 00:11:45.455 }, 00:11:45.455 "base_bdevs_list": [ 00:11:45.455 { 00:11:45.455 "name": "spare", 00:11:45.455 "uuid": "7abb63dc-21f5-5627-847b-265974781628", 00:11:45.455 "is_configured": true, 00:11:45.455 "data_offset": 0, 00:11:45.455 "data_size": 65536 00:11:45.455 }, 00:11:45.455 { 00:11:45.455 "name": "BaseBdev2", 00:11:45.455 "uuid": "b417b844-b703-58ff-98ba-17d31eb83e96", 00:11:45.455 "is_configured": true, 00:11:45.455 "data_offset": 0, 00:11:45.455 "data_size": 65536 00:11:45.455 } 00:11:45.455 ] 00:11:45.455 }' 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=312 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:45.455 "name": "raid_bdev1", 00:11:45.455 "uuid": "65beaccb-3c7b-48a9-9d0c-f2423806130b", 00:11:45.455 "strip_size_kb": 0, 00:11:45.455 "state": "online", 00:11:45.455 "raid_level": "raid1", 00:11:45.455 "superblock": false, 00:11:45.455 "num_base_bdevs": 2, 00:11:45.455 "num_base_bdevs_discovered": 2, 00:11:45.455 "num_base_bdevs_operational": 2, 00:11:45.455 "process": { 00:11:45.455 "type": "rebuild", 00:11:45.455 "target": "spare", 00:11:45.455 "progress": { 00:11:45.455 "blocks": 12288, 00:11:45.455 "percent": 18 00:11:45.455 } 00:11:45.455 }, 00:11:45.455 "base_bdevs_list": [ 00:11:45.455 { 00:11:45.455 "name": "spare", 00:11:45.455 "uuid": "7abb63dc-21f5-5627-847b-265974781628", 00:11:45.455 "is_configured": true, 00:11:45.455 "data_offset": 0, 00:11:45.455 "data_size": 65536 00:11:45.455 }, 00:11:45.455 { 00:11:45.455 "name": "BaseBdev2", 00:11:45.455 "uuid": "b417b844-b703-58ff-98ba-17d31eb83e96", 00:11:45.455 "is_configured": true, 00:11:45.455 "data_offset": 0, 00:11:45.455 "data_size": 65536 00:11:45.455 } 00:11:45.455 ] 00:11:45.455 }' 00:11:45.455 19:02:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:45.455 [2024-12-05 19:02:02.986997] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:45.715 19:02:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:45.715 19:02:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:45.715 19:02:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:45.715 19:02:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:45.715 [2024-12-05 19:02:03.112519] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:45.975 [2024-12-05 19:02:03.330648] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:11:46.235 [2024-12-05 19:02:03.536824] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:11:46.235 [2024-12-05 19:02:03.537015] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:11:46.495 133.50 IOPS, 400.50 MiB/s [2024-12-05T19:02:04.054Z] [2024-12-05 19:02:03.850435] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:11:46.495 [2024-12-05 19:02:03.850898] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:11:46.755 19:02:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:46.755 19:02:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:46.755 19:02:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:46.755 19:02:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:46.755 19:02:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:46.755 19:02:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:46.755 19:02:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:46.755 19:02:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:46.755 19:02:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:46.755 19:02:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:46.755 [2024-12-05 19:02:04.081517] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:11:46.755 19:02:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:46.755 19:02:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:46.755 "name": "raid_bdev1", 00:11:46.755 "uuid": "65beaccb-3c7b-48a9-9d0c-f2423806130b", 00:11:46.755 "strip_size_kb": 0, 00:11:46.755 "state": "online", 00:11:46.755 "raid_level": "raid1", 00:11:46.756 "superblock": false, 00:11:46.756 "num_base_bdevs": 2, 00:11:46.756 "num_base_bdevs_discovered": 2, 00:11:46.756 "num_base_bdevs_operational": 2, 00:11:46.756 "process": { 00:11:46.756 "type": "rebuild", 00:11:46.756 "target": "spare", 00:11:46.756 "progress": { 00:11:46.756 "blocks": 26624, 00:11:46.756 "percent": 40 00:11:46.756 } 00:11:46.756 }, 00:11:46.756 "base_bdevs_list": [ 00:11:46.756 { 00:11:46.756 "name": "spare", 00:11:46.756 "uuid": "7abb63dc-21f5-5627-847b-265974781628", 00:11:46.756 "is_configured": true, 00:11:46.756 "data_offset": 0, 00:11:46.756 "data_size": 65536 00:11:46.756 }, 00:11:46.756 { 00:11:46.756 "name": "BaseBdev2", 00:11:46.756 "uuid": "b417b844-b703-58ff-98ba-17d31eb83e96", 00:11:46.756 "is_configured": true, 00:11:46.756 "data_offset": 0, 00:11:46.756 "data_size": 65536 00:11:46.756 } 00:11:46.756 ] 00:11:46.756 }' 00:11:46.756 19:02:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:46.756 19:02:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:46.756 19:02:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:46.756 19:02:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:46.756 19:02:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:47.275 118.60 IOPS, 355.80 MiB/s [2024-12-05T19:02:04.834Z] [2024-12-05 19:02:04.698630] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:47.845 [2024-12-05 19:02:05.242326] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:47.845 "name": "raid_bdev1", 00:11:47.845 "uuid": "65beaccb-3c7b-48a9-9d0c-f2423806130b", 00:11:47.845 "strip_size_kb": 0, 00:11:47.845 "state": "online", 00:11:47.845 "raid_level": "raid1", 00:11:47.845 "superblock": false, 00:11:47.845 "num_base_bdevs": 2, 00:11:47.845 "num_base_bdevs_discovered": 2, 00:11:47.845 "num_base_bdevs_operational": 2, 00:11:47.845 "process": { 00:11:47.845 "type": "rebuild", 00:11:47.845 "target": "spare", 00:11:47.845 "progress": { 00:11:47.845 "blocks": 45056, 00:11:47.845 "percent": 68 00:11:47.845 } 00:11:47.845 }, 00:11:47.845 "base_bdevs_list": [ 00:11:47.845 { 00:11:47.845 "name": "spare", 00:11:47.845 "uuid": "7abb63dc-21f5-5627-847b-265974781628", 00:11:47.845 "is_configured": true, 00:11:47.845 "data_offset": 0, 00:11:47.845 "data_size": 65536 00:11:47.845 }, 00:11:47.845 { 00:11:47.845 "name": "BaseBdev2", 00:11:47.845 "uuid": "b417b844-b703-58ff-98ba-17d31eb83e96", 00:11:47.845 "is_configured": true, 00:11:47.845 "data_offset": 0, 00:11:47.845 "data_size": 65536 00:11:47.845 } 00:11:47.845 ] 00:11:47.845 }' 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:47.845 19:02:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:48.104 [2024-12-05 19:02:05.466927] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:11:48.363 105.17 IOPS, 315.50 MiB/s [2024-12-05T19:02:05.922Z] [2024-12-05 19:02:05.675134] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:11:48.932 [2024-12-05 19:02:06.331684] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:48.932 "name": "raid_bdev1", 00:11:48.932 "uuid": "65beaccb-3c7b-48a9-9d0c-f2423806130b", 00:11:48.932 "strip_size_kb": 0, 00:11:48.932 "state": "online", 00:11:48.932 "raid_level": "raid1", 00:11:48.932 "superblock": false, 00:11:48.932 "num_base_bdevs": 2, 00:11:48.932 "num_base_bdevs_discovered": 2, 00:11:48.932 "num_base_bdevs_operational": 2, 00:11:48.932 "process": { 00:11:48.932 "type": "rebuild", 00:11:48.932 "target": "spare", 00:11:48.932 "progress": { 00:11:48.932 "blocks": 65536, 00:11:48.932 "percent": 100 00:11:48.932 } 00:11:48.932 }, 00:11:48.932 "base_bdevs_list": [ 00:11:48.932 { 00:11:48.932 "name": "spare", 00:11:48.932 "uuid": "7abb63dc-21f5-5627-847b-265974781628", 00:11:48.932 "is_configured": true, 00:11:48.932 "data_offset": 0, 00:11:48.932 "data_size": 65536 00:11:48.932 }, 00:11:48.932 { 00:11:48.932 "name": "BaseBdev2", 00:11:48.932 "uuid": "b417b844-b703-58ff-98ba-17d31eb83e96", 00:11:48.932 "is_configured": true, 00:11:48.932 "data_offset": 0, 00:11:48.932 "data_size": 65536 00:11:48.932 } 00:11:48.932 ] 00:11:48.932 }' 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:48.932 [2024-12-05 19:02:06.431504] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:48.932 [2024-12-05 19:02:06.433115] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:48.932 19:02:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:50.129 95.86 IOPS, 287.57 MiB/s [2024-12-05T19:02:07.689Z] 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:50.130 "name": "raid_bdev1", 00:11:50.130 "uuid": "65beaccb-3c7b-48a9-9d0c-f2423806130b", 00:11:50.130 "strip_size_kb": 0, 00:11:50.130 "state": "online", 00:11:50.130 "raid_level": "raid1", 00:11:50.130 "superblock": false, 00:11:50.130 "num_base_bdevs": 2, 00:11:50.130 "num_base_bdevs_discovered": 2, 00:11:50.130 "num_base_bdevs_operational": 2, 00:11:50.130 "base_bdevs_list": [ 00:11:50.130 { 00:11:50.130 "name": "spare", 00:11:50.130 "uuid": "7abb63dc-21f5-5627-847b-265974781628", 00:11:50.130 "is_configured": true, 00:11:50.130 "data_offset": 0, 00:11:50.130 "data_size": 65536 00:11:50.130 }, 00:11:50.130 { 00:11:50.130 "name": "BaseBdev2", 00:11:50.130 "uuid": "b417b844-b703-58ff-98ba-17d31eb83e96", 00:11:50.130 "is_configured": true, 00:11:50.130 "data_offset": 0, 00:11:50.130 "data_size": 65536 00:11:50.130 } 00:11:50.130 ] 00:11:50.130 }' 00:11:50.130 88.12 IOPS, 264.38 MiB/s [2024-12-05T19:02:07.689Z] 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:50.130 "name": "raid_bdev1", 00:11:50.130 "uuid": "65beaccb-3c7b-48a9-9d0c-f2423806130b", 00:11:50.130 "strip_size_kb": 0, 00:11:50.130 "state": "online", 00:11:50.130 "raid_level": "raid1", 00:11:50.130 "superblock": false, 00:11:50.130 "num_base_bdevs": 2, 00:11:50.130 "num_base_bdevs_discovered": 2, 00:11:50.130 "num_base_bdevs_operational": 2, 00:11:50.130 "base_bdevs_list": [ 00:11:50.130 { 00:11:50.130 "name": "spare", 00:11:50.130 "uuid": "7abb63dc-21f5-5627-847b-265974781628", 00:11:50.130 "is_configured": true, 00:11:50.130 "data_offset": 0, 00:11:50.130 "data_size": 65536 00:11:50.130 }, 00:11:50.130 { 00:11:50.130 "name": "BaseBdev2", 00:11:50.130 "uuid": "b417b844-b703-58ff-98ba-17d31eb83e96", 00:11:50.130 "is_configured": true, 00:11:50.130 "data_offset": 0, 00:11:50.130 "data_size": 65536 00:11:50.130 } 00:11:50.130 ] 00:11:50.130 }' 00:11:50.130 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.390 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:50.390 "name": "raid_bdev1", 00:11:50.390 "uuid": "65beaccb-3c7b-48a9-9d0c-f2423806130b", 00:11:50.390 "strip_size_kb": 0, 00:11:50.390 "state": "online", 00:11:50.390 "raid_level": "raid1", 00:11:50.390 "superblock": false, 00:11:50.390 "num_base_bdevs": 2, 00:11:50.390 "num_base_bdevs_discovered": 2, 00:11:50.390 "num_base_bdevs_operational": 2, 00:11:50.390 "base_bdevs_list": [ 00:11:50.390 { 00:11:50.390 "name": "spare", 00:11:50.390 "uuid": "7abb63dc-21f5-5627-847b-265974781628", 00:11:50.390 "is_configured": true, 00:11:50.390 "data_offset": 0, 00:11:50.390 "data_size": 65536 00:11:50.390 }, 00:11:50.390 { 00:11:50.390 "name": "BaseBdev2", 00:11:50.390 "uuid": "b417b844-b703-58ff-98ba-17d31eb83e96", 00:11:50.390 "is_configured": true, 00:11:50.391 "data_offset": 0, 00:11:50.391 "data_size": 65536 00:11:50.391 } 00:11:50.391 ] 00:11:50.391 }' 00:11:50.391 19:02:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:50.391 19:02:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.651 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:50.651 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.651 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.651 [2024-12-05 19:02:08.149387] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:50.651 [2024-12-05 19:02:08.149493] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:50.911 00:11:50.911 Latency(us) 00:11:50.911 [2024-12-05T19:02:08.470Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:50.911 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:11:50.911 raid_bdev1 : 8.72 83.13 249.38 0.00 0.00 16956.54 271.87 107604.96 00:11:50.911 [2024-12-05T19:02:08.470Z] =================================================================================================================== 00:11:50.911 [2024-12-05T19:02:08.470Z] Total : 83.13 249.38 0.00 0.00 16956.54 271.87 107604.96 00:11:50.911 [2024-12-05 19:02:08.248444] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:50.911 [2024-12-05 19:02:08.248543] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:50.911 [2024-12-05 19:02:08.248650] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:50.911 [2024-12-05 19:02:08.248726] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:50.911 { 00:11:50.911 "results": [ 00:11:50.911 { 00:11:50.911 "job": "raid_bdev1", 00:11:50.911 "core_mask": "0x1", 00:11:50.911 "workload": "randrw", 00:11:50.911 "percentage": 50, 00:11:50.911 "status": "finished", 00:11:50.911 "queue_depth": 2, 00:11:50.911 "io_size": 3145728, 00:11:50.911 "runtime": 8.721593, 00:11:50.911 "iops": 83.12701590179684, 00:11:50.911 "mibps": 249.38104770539053, 00:11:50.911 "io_failed": 0, 00:11:50.911 "io_timeout": 0, 00:11:50.911 "avg_latency_us": 16956.54444330673, 00:11:50.911 "min_latency_us": 271.87423580786026, 00:11:50.911 "max_latency_us": 107604.96069868996 00:11:50.911 } 00:11:50.911 ], 00:11:50.911 "core_count": 1 00:11:50.911 } 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:50.911 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:11:51.172 /dev/nbd0 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:51.172 1+0 records in 00:11:51.172 1+0 records out 00:11:51.172 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000426385 s, 9.6 MB/s 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:51.172 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:11:51.432 /dev/nbd1 00:11:51.432 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:51.432 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:51.432 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:11:51.432 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:11:51.432 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:11:51.432 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:11:51.432 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:11:51.432 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:11:51.432 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:11:51.432 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:51.433 1+0 records in 00:11:51.433 1+0 records out 00:11:51.433 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000340748 s, 12.0 MB/s 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:51.433 19:02:08 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:51.693 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:51.694 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:51.694 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:51.694 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:51.694 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:51.694 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:51.694 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:11:51.694 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:51.694 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:51.694 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:51.694 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:51.694 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:51.694 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:11:51.694 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:51.694 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:51.694 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 85673 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # '[' -z 85673 ']' 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # kill -0 85673 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # uname 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85673 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85673' 00:11:51.963 killing process with pid 85673 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@973 -- # kill 85673 00:11:51.963 Received shutdown signal, test time was about 9.784187 seconds 00:11:51.963 00:11:51.963 Latency(us) 00:11:51.963 [2024-12-05T19:02:09.522Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:51.963 [2024-12-05T19:02:09.522Z] =================================================================================================================== 00:11:51.963 [2024-12-05T19:02:09.522Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:11:51.963 [2024-12-05 19:02:09.305781] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:51.963 19:02:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@978 -- # wait 85673 00:11:51.963 [2024-12-05 19:02:09.331781] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:11:52.268 00:11:52.268 real 0m11.641s 00:11:52.268 user 0m14.755s 00:11:52.268 sys 0m1.362s 00:11:52.268 ************************************ 00:11:52.268 END TEST raid_rebuild_test_io 00:11:52.268 ************************************ 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.268 19:02:09 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 2 true true true 00:11:52.268 19:02:09 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:11:52.268 19:02:09 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:52.268 19:02:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:52.268 ************************************ 00:11:52.268 START TEST raid_rebuild_test_sb_io 00:11:52.268 ************************************ 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true true true 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=85989 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 85989 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # '[' -z 85989 ']' 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:52.268 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:52.268 19:02:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.268 [2024-12-05 19:02:09.711110] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:11:52.268 [2024-12-05 19:02:09.711338] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:11:52.268 Zero copy mechanism will not be used. 00:11:52.268 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85989 ] 00:11:52.528 [2024-12-05 19:02:09.867332] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:52.528 [2024-12-05 19:02:09.892908] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:52.528 [2024-12-05 19:02:09.937665] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:52.528 [2024-12-05 19:02:09.937799] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # return 0 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.099 BaseBdev1_malloc 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.099 [2024-12-05 19:02:10.535294] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:53.099 [2024-12-05 19:02:10.535364] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:53.099 [2024-12-05 19:02:10.535401] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:53.099 [2024-12-05 19:02:10.535415] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:53.099 [2024-12-05 19:02:10.537523] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:53.099 [2024-12-05 19:02:10.537616] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:53.099 BaseBdev1 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.099 BaseBdev2_malloc 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.099 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.100 [2024-12-05 19:02:10.560099] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:53.100 [2024-12-05 19:02:10.560159] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:53.100 [2024-12-05 19:02:10.560185] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:53.100 [2024-12-05 19:02:10.560197] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:53.100 [2024-12-05 19:02:10.562219] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:53.100 [2024-12-05 19:02:10.562265] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:53.100 BaseBdev2 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.100 spare_malloc 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.100 spare_delay 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.100 [2024-12-05 19:02:10.601016] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:53.100 [2024-12-05 19:02:10.601142] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:53.100 [2024-12-05 19:02:10.601168] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:53.100 [2024-12-05 19:02:10.601178] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:53.100 [2024-12-05 19:02:10.603260] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:53.100 [2024-12-05 19:02:10.603303] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:53.100 spare 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.100 [2024-12-05 19:02:10.613044] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:53.100 [2024-12-05 19:02:10.614868] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:53.100 [2024-12-05 19:02:10.615040] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:53.100 [2024-12-05 19:02:10.615054] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:53.100 [2024-12-05 19:02:10.615351] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:11:53.100 [2024-12-05 19:02:10.615508] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:53.100 [2024-12-05 19:02:10.615522] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:53.100 [2024-12-05 19:02:10.615647] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.100 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.361 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:53.361 "name": "raid_bdev1", 00:11:53.361 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:11:53.361 "strip_size_kb": 0, 00:11:53.361 "state": "online", 00:11:53.361 "raid_level": "raid1", 00:11:53.361 "superblock": true, 00:11:53.361 "num_base_bdevs": 2, 00:11:53.361 "num_base_bdevs_discovered": 2, 00:11:53.361 "num_base_bdevs_operational": 2, 00:11:53.361 "base_bdevs_list": [ 00:11:53.361 { 00:11:53.361 "name": "BaseBdev1", 00:11:53.361 "uuid": "7c131c63-64a4-560f-b313-52aece266e57", 00:11:53.361 "is_configured": true, 00:11:53.361 "data_offset": 2048, 00:11:53.361 "data_size": 63488 00:11:53.361 }, 00:11:53.361 { 00:11:53.361 "name": "BaseBdev2", 00:11:53.361 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:11:53.361 "is_configured": true, 00:11:53.361 "data_offset": 2048, 00:11:53.361 "data_size": 63488 00:11:53.361 } 00:11:53.361 ] 00:11:53.361 }' 00:11:53.361 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:53.361 19:02:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.622 [2024-12-05 19:02:11.020592] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.622 [2024-12-05 19:02:11.116182] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:53.622 "name": "raid_bdev1", 00:11:53.622 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:11:53.622 "strip_size_kb": 0, 00:11:53.622 "state": "online", 00:11:53.622 "raid_level": "raid1", 00:11:53.622 "superblock": true, 00:11:53.622 "num_base_bdevs": 2, 00:11:53.622 "num_base_bdevs_discovered": 1, 00:11:53.622 "num_base_bdevs_operational": 1, 00:11:53.622 "base_bdevs_list": [ 00:11:53.622 { 00:11:53.622 "name": null, 00:11:53.622 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:53.622 "is_configured": false, 00:11:53.622 "data_offset": 0, 00:11:53.622 "data_size": 63488 00:11:53.622 }, 00:11:53.622 { 00:11:53.622 "name": "BaseBdev2", 00:11:53.622 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:11:53.622 "is_configured": true, 00:11:53.622 "data_offset": 2048, 00:11:53.622 "data_size": 63488 00:11:53.622 } 00:11:53.622 ] 00:11:53.622 }' 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:53.622 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.882 [2024-12-05 19:02:11.223055] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:11:53.882 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:53.882 Zero copy mechanism will not be used. 00:11:53.882 Running I/O for 60 seconds... 00:11:54.142 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:54.142 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:54.142 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:54.142 [2024-12-05 19:02:11.532378] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:54.142 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:54.142 19:02:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:54.142 [2024-12-05 19:02:11.571175] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:11:54.142 [2024-12-05 19:02:11.573256] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:54.142 [2024-12-05 19:02:11.697091] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:54.142 [2024-12-05 19:02:11.697536] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:54.402 [2024-12-05 19:02:11.820814] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:54.402 [2024-12-05 19:02:11.821028] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:54.973 181.00 IOPS, 543.00 MiB/s [2024-12-05T19:02:12.532Z] [2024-12-05 19:02:12.261260] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:54.973 [2024-12-05 19:02:12.261684] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.233 [2024-12-05 19:02:12.588417] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:55.233 "name": "raid_bdev1", 00:11:55.233 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:11:55.233 "strip_size_kb": 0, 00:11:55.233 "state": "online", 00:11:55.233 "raid_level": "raid1", 00:11:55.233 "superblock": true, 00:11:55.233 "num_base_bdevs": 2, 00:11:55.233 "num_base_bdevs_discovered": 2, 00:11:55.233 "num_base_bdevs_operational": 2, 00:11:55.233 "process": { 00:11:55.233 "type": "rebuild", 00:11:55.233 "target": "spare", 00:11:55.233 "progress": { 00:11:55.233 "blocks": 12288, 00:11:55.233 "percent": 19 00:11:55.233 } 00:11:55.233 }, 00:11:55.233 "base_bdevs_list": [ 00:11:55.233 { 00:11:55.233 "name": "spare", 00:11:55.233 "uuid": "b83547a1-14e2-5336-82ea-a70f91c7c22b", 00:11:55.233 "is_configured": true, 00:11:55.233 "data_offset": 2048, 00:11:55.233 "data_size": 63488 00:11:55.233 }, 00:11:55.233 { 00:11:55.233 "name": "BaseBdev2", 00:11:55.233 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:11:55.233 "is_configured": true, 00:11:55.233 "data_offset": 2048, 00:11:55.233 "data_size": 63488 00:11:55.233 } 00:11:55.233 ] 00:11:55.233 }' 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:55.233 [2024-12-05 19:02:12.694616] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:55.233 [2024-12-05 19:02:12.702109] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:55.233 [2024-12-05 19:02:12.702373] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:55.233 [2024-12-05 19:02:12.708629] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:55.233 [2024-12-05 19:02:12.710663] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:55.233 [2024-12-05 19:02:12.710712] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:55.233 [2024-12-05 19:02:12.710740] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:55.233 [2024-12-05 19:02:12.732568] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000026d0 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:55.233 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.493 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:55.493 "name": "raid_bdev1", 00:11:55.493 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:11:55.493 "strip_size_kb": 0, 00:11:55.493 "state": "online", 00:11:55.493 "raid_level": "raid1", 00:11:55.493 "superblock": true, 00:11:55.493 "num_base_bdevs": 2, 00:11:55.493 "num_base_bdevs_discovered": 1, 00:11:55.493 "num_base_bdevs_operational": 1, 00:11:55.493 "base_bdevs_list": [ 00:11:55.493 { 00:11:55.493 "name": null, 00:11:55.493 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:55.493 "is_configured": false, 00:11:55.493 "data_offset": 0, 00:11:55.493 "data_size": 63488 00:11:55.493 }, 00:11:55.493 { 00:11:55.493 "name": "BaseBdev2", 00:11:55.493 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:11:55.493 "is_configured": true, 00:11:55.493 "data_offset": 2048, 00:11:55.493 "data_size": 63488 00:11:55.493 } 00:11:55.493 ] 00:11:55.493 }' 00:11:55.493 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:55.493 19:02:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:55.752 "name": "raid_bdev1", 00:11:55.752 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:11:55.752 "strip_size_kb": 0, 00:11:55.752 "state": "online", 00:11:55.752 "raid_level": "raid1", 00:11:55.752 "superblock": true, 00:11:55.752 "num_base_bdevs": 2, 00:11:55.752 "num_base_bdevs_discovered": 1, 00:11:55.752 "num_base_bdevs_operational": 1, 00:11:55.752 "base_bdevs_list": [ 00:11:55.752 { 00:11:55.752 "name": null, 00:11:55.752 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:55.752 "is_configured": false, 00:11:55.752 "data_offset": 0, 00:11:55.752 "data_size": 63488 00:11:55.752 }, 00:11:55.752 { 00:11:55.752 "name": "BaseBdev2", 00:11:55.752 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:11:55.752 "is_configured": true, 00:11:55.752 "data_offset": 2048, 00:11:55.752 "data_size": 63488 00:11:55.752 } 00:11:55.752 ] 00:11:55.752 }' 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:55.752 207.00 IOPS, 621.00 MiB/s [2024-12-05T19:02:13.311Z] 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:55.752 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:56.011 [2024-12-05 19:02:13.312791] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:56.011 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:56.011 19:02:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:56.011 [2024-12-05 19:02:13.364916] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:11:56.011 [2024-12-05 19:02:13.366963] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:56.011 [2024-12-05 19:02:13.479134] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:56.011 [2024-12-05 19:02:13.479475] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:56.270 [2024-12-05 19:02:13.601409] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:56.270 [2024-12-05 19:02:13.601553] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:56.529 [2024-12-05 19:02:13.935858] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:56.788 [2024-12-05 19:02:14.160060] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:57.047 175.67 IOPS, 527.00 MiB/s [2024-12-05T19:02:14.606Z] 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:57.047 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:57.047 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:57.047 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:57.047 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:57.047 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:57.047 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:57.047 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:57.047 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:57.047 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:57.047 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:57.047 "name": "raid_bdev1", 00:11:57.047 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:11:57.047 "strip_size_kb": 0, 00:11:57.047 "state": "online", 00:11:57.047 "raid_level": "raid1", 00:11:57.047 "superblock": true, 00:11:57.047 "num_base_bdevs": 2, 00:11:57.047 "num_base_bdevs_discovered": 2, 00:11:57.047 "num_base_bdevs_operational": 2, 00:11:57.047 "process": { 00:11:57.047 "type": "rebuild", 00:11:57.047 "target": "spare", 00:11:57.047 "progress": { 00:11:57.047 "blocks": 12288, 00:11:57.047 "percent": 19 00:11:57.047 } 00:11:57.047 }, 00:11:57.047 "base_bdevs_list": [ 00:11:57.047 { 00:11:57.047 "name": "spare", 00:11:57.047 "uuid": "b83547a1-14e2-5336-82ea-a70f91c7c22b", 00:11:57.047 "is_configured": true, 00:11:57.047 "data_offset": 2048, 00:11:57.047 "data_size": 63488 00:11:57.047 }, 00:11:57.047 { 00:11:57.047 "name": "BaseBdev2", 00:11:57.047 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:11:57.047 "is_configured": true, 00:11:57.047 "data_offset": 2048, 00:11:57.047 "data_size": 63488 00:11:57.047 } 00:11:57.047 ] 00:11:57.047 }' 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:11:57.048 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=324 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:57.048 [2024-12-05 19:02:14.521734] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:57.048 "name": "raid_bdev1", 00:11:57.048 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:11:57.048 "strip_size_kb": 0, 00:11:57.048 "state": "online", 00:11:57.048 "raid_level": "raid1", 00:11:57.048 "superblock": true, 00:11:57.048 "num_base_bdevs": 2, 00:11:57.048 "num_base_bdevs_discovered": 2, 00:11:57.048 "num_base_bdevs_operational": 2, 00:11:57.048 "process": { 00:11:57.048 "type": "rebuild", 00:11:57.048 "target": "spare", 00:11:57.048 "progress": { 00:11:57.048 "blocks": 14336, 00:11:57.048 "percent": 22 00:11:57.048 } 00:11:57.048 }, 00:11:57.048 "base_bdevs_list": [ 00:11:57.048 { 00:11:57.048 "name": "spare", 00:11:57.048 "uuid": "b83547a1-14e2-5336-82ea-a70f91c7c22b", 00:11:57.048 "is_configured": true, 00:11:57.048 "data_offset": 2048, 00:11:57.048 "data_size": 63488 00:11:57.048 }, 00:11:57.048 { 00:11:57.048 "name": "BaseBdev2", 00:11:57.048 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:11:57.048 "is_configured": true, 00:11:57.048 "data_offset": 2048, 00:11:57.048 "data_size": 63488 00:11:57.048 } 00:11:57.048 ] 00:11:57.048 }' 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:57.048 19:02:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:57.615 [2024-12-05 19:02:15.172670] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:11:57.873 [2024-12-05 19:02:15.173105] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:11:58.131 140.25 IOPS, 420.75 MiB/s [2024-12-05T19:02:15.690Z] 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:58.131 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:58.131 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:58.131 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:58.131 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:58.131 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:58.131 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.131 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:58.131 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:58.131 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:58.131 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:58.131 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:58.131 "name": "raid_bdev1", 00:11:58.131 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:11:58.131 "strip_size_kb": 0, 00:11:58.131 "state": "online", 00:11:58.131 "raid_level": "raid1", 00:11:58.131 "superblock": true, 00:11:58.131 "num_base_bdevs": 2, 00:11:58.131 "num_base_bdevs_discovered": 2, 00:11:58.131 "num_base_bdevs_operational": 2, 00:11:58.131 "process": { 00:11:58.131 "type": "rebuild", 00:11:58.131 "target": "spare", 00:11:58.131 "progress": { 00:11:58.131 "blocks": 32768, 00:11:58.131 "percent": 51 00:11:58.131 } 00:11:58.131 }, 00:11:58.131 "base_bdevs_list": [ 00:11:58.131 { 00:11:58.131 "name": "spare", 00:11:58.131 "uuid": "b83547a1-14e2-5336-82ea-a70f91c7c22b", 00:11:58.131 "is_configured": true, 00:11:58.131 "data_offset": 2048, 00:11:58.131 "data_size": 63488 00:11:58.131 }, 00:11:58.131 { 00:11:58.131 "name": "BaseBdev2", 00:11:58.131 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:11:58.131 "is_configured": true, 00:11:58.131 "data_offset": 2048, 00:11:58.131 "data_size": 63488 00:11:58.131 } 00:11:58.131 ] 00:11:58.131 }' 00:11:58.131 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:58.388 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:58.388 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:58.388 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:58.388 19:02:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:58.957 125.20 IOPS, 375.60 MiB/s [2024-12-05T19:02:16.516Z] [2024-12-05 19:02:16.510801] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:11:59.218 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:59.218 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:59.218 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:59.218 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:59.218 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:59.218 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:59.218 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:59.218 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:59.218 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:59.218 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:59.478 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:59.478 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:59.478 "name": "raid_bdev1", 00:11:59.478 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:11:59.478 "strip_size_kb": 0, 00:11:59.478 "state": "online", 00:11:59.478 "raid_level": "raid1", 00:11:59.478 "superblock": true, 00:11:59.478 "num_base_bdevs": 2, 00:11:59.478 "num_base_bdevs_discovered": 2, 00:11:59.478 "num_base_bdevs_operational": 2, 00:11:59.478 "process": { 00:11:59.478 "type": "rebuild", 00:11:59.478 "target": "spare", 00:11:59.478 "progress": { 00:11:59.478 "blocks": 55296, 00:11:59.478 "percent": 87 00:11:59.478 } 00:11:59.478 }, 00:11:59.478 "base_bdevs_list": [ 00:11:59.478 { 00:11:59.478 "name": "spare", 00:11:59.478 "uuid": "b83547a1-14e2-5336-82ea-a70f91c7c22b", 00:11:59.478 "is_configured": true, 00:11:59.478 "data_offset": 2048, 00:11:59.478 "data_size": 63488 00:11:59.478 }, 00:11:59.478 { 00:11:59.478 "name": "BaseBdev2", 00:11:59.478 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:11:59.478 "is_configured": true, 00:11:59.478 "data_offset": 2048, 00:11:59.478 "data_size": 63488 00:11:59.478 } 00:11:59.478 ] 00:11:59.478 }' 00:11:59.478 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:59.478 [2024-12-05 19:02:16.838489] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:11:59.478 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:59.478 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:59.478 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:59.478 19:02:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:59.738 [2024-12-05 19:02:17.164640] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:59.738 111.67 IOPS, 335.00 MiB/s [2024-12-05T19:02:17.297Z] [2024-12-05 19:02:17.269399] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:59.738 [2024-12-05 19:02:17.271231] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:00.680 19:02:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:00.680 19:02:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:00.680 19:02:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:00.680 19:02:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:00.680 19:02:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:00.680 19:02:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:00.680 19:02:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.680 19:02:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:00.680 19:02:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.680 19:02:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:00.680 19:02:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.680 19:02:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:00.680 "name": "raid_bdev1", 00:12:00.680 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:12:00.680 "strip_size_kb": 0, 00:12:00.680 "state": "online", 00:12:00.680 "raid_level": "raid1", 00:12:00.680 "superblock": true, 00:12:00.680 "num_base_bdevs": 2, 00:12:00.680 "num_base_bdevs_discovered": 2, 00:12:00.680 "num_base_bdevs_operational": 2, 00:12:00.680 "base_bdevs_list": [ 00:12:00.680 { 00:12:00.680 "name": "spare", 00:12:00.680 "uuid": "b83547a1-14e2-5336-82ea-a70f91c7c22b", 00:12:00.680 "is_configured": true, 00:12:00.680 "data_offset": 2048, 00:12:00.680 "data_size": 63488 00:12:00.680 }, 00:12:00.680 { 00:12:00.680 "name": "BaseBdev2", 00:12:00.680 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:12:00.680 "is_configured": true, 00:12:00.680 "data_offset": 2048, 00:12:00.680 "data_size": 63488 00:12:00.680 } 00:12:00.680 ] 00:12:00.680 }' 00:12:00.680 19:02:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:00.680 "name": "raid_bdev1", 00:12:00.680 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:12:00.680 "strip_size_kb": 0, 00:12:00.680 "state": "online", 00:12:00.680 "raid_level": "raid1", 00:12:00.680 "superblock": true, 00:12:00.680 "num_base_bdevs": 2, 00:12:00.680 "num_base_bdevs_discovered": 2, 00:12:00.680 "num_base_bdevs_operational": 2, 00:12:00.680 "base_bdevs_list": [ 00:12:00.680 { 00:12:00.680 "name": "spare", 00:12:00.680 "uuid": "b83547a1-14e2-5336-82ea-a70f91c7c22b", 00:12:00.680 "is_configured": true, 00:12:00.680 "data_offset": 2048, 00:12:00.680 "data_size": 63488 00:12:00.680 }, 00:12:00.680 { 00:12:00.680 "name": "BaseBdev2", 00:12:00.680 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:12:00.680 "is_configured": true, 00:12:00.680 "data_offset": 2048, 00:12:00.680 "data_size": 63488 00:12:00.680 } 00:12:00.680 ] 00:12:00.680 }' 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:00.680 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:00.680 "name": "raid_bdev1", 00:12:00.681 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:12:00.681 "strip_size_kb": 0, 00:12:00.681 "state": "online", 00:12:00.681 "raid_level": "raid1", 00:12:00.681 "superblock": true, 00:12:00.681 "num_base_bdevs": 2, 00:12:00.681 "num_base_bdevs_discovered": 2, 00:12:00.681 "num_base_bdevs_operational": 2, 00:12:00.681 "base_bdevs_list": [ 00:12:00.681 { 00:12:00.681 "name": "spare", 00:12:00.681 "uuid": "b83547a1-14e2-5336-82ea-a70f91c7c22b", 00:12:00.681 "is_configured": true, 00:12:00.681 "data_offset": 2048, 00:12:00.681 "data_size": 63488 00:12:00.681 }, 00:12:00.681 { 00:12:00.681 "name": "BaseBdev2", 00:12:00.681 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:12:00.681 "is_configured": true, 00:12:00.681 "data_offset": 2048, 00:12:00.681 "data_size": 63488 00:12:00.681 } 00:12:00.681 ] 00:12:00.681 }' 00:12:00.681 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:00.681 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.254 100.71 IOPS, 302.14 MiB/s [2024-12-05T19:02:18.813Z] 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:01.254 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.254 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.254 [2024-12-05 19:02:18.646332] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:01.254 [2024-12-05 19:02:18.646435] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:01.254 00:12:01.254 Latency(us) 00:12:01.254 [2024-12-05T19:02:18.813Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:01.254 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:12:01.254 raid_bdev1 : 7.45 97.09 291.28 0.00 0.00 14603.64 273.66 107147.07 00:12:01.254 [2024-12-05T19:02:18.813Z] =================================================================================================================== 00:12:01.254 [2024-12-05T19:02:18.813Z] Total : 97.09 291.28 0.00 0.00 14603.64 273.66 107147.07 00:12:01.254 [2024-12-05 19:02:18.661330] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:01.254 [2024-12-05 19:02:18.661426] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:01.254 [2024-12-05 19:02:18.661538] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:01.254 [2024-12-05 19:02:18.661595] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:01.254 { 00:12:01.254 "results": [ 00:12:01.254 { 00:12:01.254 "job": "raid_bdev1", 00:12:01.254 "core_mask": "0x1", 00:12:01.254 "workload": "randrw", 00:12:01.254 "percentage": 50, 00:12:01.254 "status": "finished", 00:12:01.254 "queue_depth": 2, 00:12:01.254 "io_size": 3145728, 00:12:01.254 "runtime": 7.446538, 00:12:01.254 "iops": 97.09209836839617, 00:12:01.254 "mibps": 291.2762951051885, 00:12:01.254 "io_failed": 0, 00:12:01.254 "io_timeout": 0, 00:12:01.254 "avg_latency_us": 14603.637642766977, 00:12:01.254 "min_latency_us": 273.6628820960699, 00:12:01.254 "max_latency_us": 107147.0672489083 00:12:01.254 } 00:12:01.254 ], 00:12:01.254 "core_count": 1 00:12:01.254 } 00:12:01.254 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.254 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.254 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:12:01.254 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:01.254 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:01.254 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:01.255 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:01.255 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:01.255 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:12:01.255 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:12:01.255 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:01.255 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:12:01.255 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:01.255 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:01.255 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:01.255 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:01.255 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:01.255 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:01.255 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:12:01.515 /dev/nbd0 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:01.515 1+0 records in 00:12:01.515 1+0 records out 00:12:01.515 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000557886 s, 7.3 MB/s 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:01.515 19:02:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:12:01.776 /dev/nbd1 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:01.776 1+0 records in 00:12:01.776 1+0 records out 00:12:01.776 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000539446 s, 7.6 MB/s 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:01.776 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:01.777 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:01.777 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:01.777 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:02.038 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:02.038 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:02.038 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:02.038 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:02.038 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:02.038 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:02.038 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:02.038 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:02.038 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:02.038 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:02.038 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:02.038 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:02.038 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:02.038 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:02.038 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:02.298 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:02.299 [2024-12-05 19:02:19.702328] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:02.299 [2024-12-05 19:02:19.702396] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:02.299 [2024-12-05 19:02:19.702419] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:12:02.299 [2024-12-05 19:02:19.702432] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:02.299 [2024-12-05 19:02:19.704693] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:02.299 [2024-12-05 19:02:19.704784] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:02.299 [2024-12-05 19:02:19.704902] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:02.299 [2024-12-05 19:02:19.704960] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:02.299 [2024-12-05 19:02:19.705093] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:02.299 spare 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:02.299 [2024-12-05 19:02:19.804995] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:12:02.299 [2024-12-05 19:02:19.805030] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:02.299 [2024-12-05 19:02:19.805304] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027720 00:12:02.299 [2024-12-05 19:02:19.805437] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:12:02.299 [2024-12-05 19:02:19.805451] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:12:02.299 [2024-12-05 19:02:19.805583] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:02.299 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.559 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:02.559 "name": "raid_bdev1", 00:12:02.559 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:12:02.559 "strip_size_kb": 0, 00:12:02.559 "state": "online", 00:12:02.559 "raid_level": "raid1", 00:12:02.559 "superblock": true, 00:12:02.559 "num_base_bdevs": 2, 00:12:02.559 "num_base_bdevs_discovered": 2, 00:12:02.559 "num_base_bdevs_operational": 2, 00:12:02.559 "base_bdevs_list": [ 00:12:02.559 { 00:12:02.559 "name": "spare", 00:12:02.559 "uuid": "b83547a1-14e2-5336-82ea-a70f91c7c22b", 00:12:02.559 "is_configured": true, 00:12:02.559 "data_offset": 2048, 00:12:02.559 "data_size": 63488 00:12:02.559 }, 00:12:02.559 { 00:12:02.559 "name": "BaseBdev2", 00:12:02.559 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:12:02.559 "is_configured": true, 00:12:02.559 "data_offset": 2048, 00:12:02.559 "data_size": 63488 00:12:02.559 } 00:12:02.559 ] 00:12:02.559 }' 00:12:02.559 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:02.559 19:02:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:02.819 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:02.819 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:02.819 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:02.819 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:02.819 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:02.819 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:02.819 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:02.819 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:02.819 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:02.819 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:02.819 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:02.819 "name": "raid_bdev1", 00:12:02.819 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:12:02.819 "strip_size_kb": 0, 00:12:02.819 "state": "online", 00:12:02.819 "raid_level": "raid1", 00:12:02.819 "superblock": true, 00:12:02.819 "num_base_bdevs": 2, 00:12:02.819 "num_base_bdevs_discovered": 2, 00:12:02.819 "num_base_bdevs_operational": 2, 00:12:02.819 "base_bdevs_list": [ 00:12:02.819 { 00:12:02.819 "name": "spare", 00:12:02.819 "uuid": "b83547a1-14e2-5336-82ea-a70f91c7c22b", 00:12:02.819 "is_configured": true, 00:12:02.819 "data_offset": 2048, 00:12:02.819 "data_size": 63488 00:12:02.819 }, 00:12:02.819 { 00:12:02.819 "name": "BaseBdev2", 00:12:02.819 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:12:02.819 "is_configured": true, 00:12:02.819 "data_offset": 2048, 00:12:02.819 "data_size": 63488 00:12:02.819 } 00:12:02.819 ] 00:12:02.819 }' 00:12:02.819 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:02.819 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:02.819 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.079 [2024-12-05 19:02:20.453168] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:03.079 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.080 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.080 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.080 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.080 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:03.080 "name": "raid_bdev1", 00:12:03.080 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:12:03.080 "strip_size_kb": 0, 00:12:03.080 "state": "online", 00:12:03.080 "raid_level": "raid1", 00:12:03.080 "superblock": true, 00:12:03.080 "num_base_bdevs": 2, 00:12:03.080 "num_base_bdevs_discovered": 1, 00:12:03.080 "num_base_bdevs_operational": 1, 00:12:03.080 "base_bdevs_list": [ 00:12:03.080 { 00:12:03.080 "name": null, 00:12:03.080 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:03.080 "is_configured": false, 00:12:03.080 "data_offset": 0, 00:12:03.080 "data_size": 63488 00:12:03.080 }, 00:12:03.080 { 00:12:03.080 "name": "BaseBdev2", 00:12:03.080 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:12:03.080 "is_configured": true, 00:12:03.080 "data_offset": 2048, 00:12:03.080 "data_size": 63488 00:12:03.080 } 00:12:03.080 ] 00:12:03.080 }' 00:12:03.080 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:03.080 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.340 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:03.340 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:03.340 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:03.340 [2024-12-05 19:02:20.876488] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:03.340 [2024-12-05 19:02:20.876669] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:03.340 [2024-12-05 19:02:20.876700] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:03.340 [2024-12-05 19:02:20.876756] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:03.340 [2024-12-05 19:02:20.882132] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000277f0 00:12:03.340 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:03.340 19:02:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:03.340 [2024-12-05 19:02:20.884029] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:04.723 19:02:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:04.723 19:02:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:04.723 19:02:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:04.723 19:02:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:04.723 19:02:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:04.723 19:02:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:04.723 19:02:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:04.723 19:02:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.723 19:02:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:04.724 19:02:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.724 19:02:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:04.724 "name": "raid_bdev1", 00:12:04.724 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:12:04.724 "strip_size_kb": 0, 00:12:04.724 "state": "online", 00:12:04.724 "raid_level": "raid1", 00:12:04.724 "superblock": true, 00:12:04.724 "num_base_bdevs": 2, 00:12:04.724 "num_base_bdevs_discovered": 2, 00:12:04.724 "num_base_bdevs_operational": 2, 00:12:04.724 "process": { 00:12:04.724 "type": "rebuild", 00:12:04.724 "target": "spare", 00:12:04.724 "progress": { 00:12:04.724 "blocks": 20480, 00:12:04.724 "percent": 32 00:12:04.724 } 00:12:04.724 }, 00:12:04.724 "base_bdevs_list": [ 00:12:04.724 { 00:12:04.724 "name": "spare", 00:12:04.724 "uuid": "b83547a1-14e2-5336-82ea-a70f91c7c22b", 00:12:04.724 "is_configured": true, 00:12:04.724 "data_offset": 2048, 00:12:04.724 "data_size": 63488 00:12:04.724 }, 00:12:04.724 { 00:12:04.724 "name": "BaseBdev2", 00:12:04.724 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:12:04.724 "is_configured": true, 00:12:04.724 "data_offset": 2048, 00:12:04.724 "data_size": 63488 00:12:04.724 } 00:12:04.724 ] 00:12:04.724 }' 00:12:04.724 19:02:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:04.724 19:02:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:04.724 19:02:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:04.724 [2024-12-05 19:02:22.044248] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:04.724 [2024-12-05 19:02:22.088103] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:04.724 [2024-12-05 19:02:22.088161] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:04.724 [2024-12-05 19:02:22.088198] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:04.724 [2024-12-05 19:02:22.088206] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:04.724 "name": "raid_bdev1", 00:12:04.724 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:12:04.724 "strip_size_kb": 0, 00:12:04.724 "state": "online", 00:12:04.724 "raid_level": "raid1", 00:12:04.724 "superblock": true, 00:12:04.724 "num_base_bdevs": 2, 00:12:04.724 "num_base_bdevs_discovered": 1, 00:12:04.724 "num_base_bdevs_operational": 1, 00:12:04.724 "base_bdevs_list": [ 00:12:04.724 { 00:12:04.724 "name": null, 00:12:04.724 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:04.724 "is_configured": false, 00:12:04.724 "data_offset": 0, 00:12:04.724 "data_size": 63488 00:12:04.724 }, 00:12:04.724 { 00:12:04.724 "name": "BaseBdev2", 00:12:04.724 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:12:04.724 "is_configured": true, 00:12:04.724 "data_offset": 2048, 00:12:04.724 "data_size": 63488 00:12:04.724 } 00:12:04.724 ] 00:12:04.724 }' 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:04.724 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:05.294 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:05.294 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:05.294 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:05.294 [2024-12-05 19:02:22.552181] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:05.294 [2024-12-05 19:02:22.552305] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:05.294 [2024-12-05 19:02:22.552354] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:12:05.294 [2024-12-05 19:02:22.552391] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:05.294 [2024-12-05 19:02:22.552883] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:05.294 [2024-12-05 19:02:22.552957] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:05.294 [2024-12-05 19:02:22.553094] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:05.294 [2024-12-05 19:02:22.553140] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:05.294 [2024-12-05 19:02:22.553191] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:05.294 [2024-12-05 19:02:22.553243] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:05.294 [2024-12-05 19:02:22.558542] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000278c0 00:12:05.294 spare 00:12:05.294 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:05.294 19:02:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:12:05.294 [2024-12-05 19:02:22.560504] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:06.238 "name": "raid_bdev1", 00:12:06.238 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:12:06.238 "strip_size_kb": 0, 00:12:06.238 "state": "online", 00:12:06.238 "raid_level": "raid1", 00:12:06.238 "superblock": true, 00:12:06.238 "num_base_bdevs": 2, 00:12:06.238 "num_base_bdevs_discovered": 2, 00:12:06.238 "num_base_bdevs_operational": 2, 00:12:06.238 "process": { 00:12:06.238 "type": "rebuild", 00:12:06.238 "target": "spare", 00:12:06.238 "progress": { 00:12:06.238 "blocks": 20480, 00:12:06.238 "percent": 32 00:12:06.238 } 00:12:06.238 }, 00:12:06.238 "base_bdevs_list": [ 00:12:06.238 { 00:12:06.238 "name": "spare", 00:12:06.238 "uuid": "b83547a1-14e2-5336-82ea-a70f91c7c22b", 00:12:06.238 "is_configured": true, 00:12:06.238 "data_offset": 2048, 00:12:06.238 "data_size": 63488 00:12:06.238 }, 00:12:06.238 { 00:12:06.238 "name": "BaseBdev2", 00:12:06.238 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:12:06.238 "is_configured": true, 00:12:06.238 "data_offset": 2048, 00:12:06.238 "data_size": 63488 00:12:06.238 } 00:12:06.238 ] 00:12:06.238 }' 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:06.238 [2024-12-05 19:02:23.720865] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:06.238 [2024-12-05 19:02:23.764714] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:06.238 [2024-12-05 19:02:23.764780] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:06.238 [2024-12-05 19:02:23.764797] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:06.238 [2024-12-05 19:02:23.764808] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.238 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:06.498 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.498 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:06.498 "name": "raid_bdev1", 00:12:06.498 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:12:06.498 "strip_size_kb": 0, 00:12:06.498 "state": "online", 00:12:06.498 "raid_level": "raid1", 00:12:06.498 "superblock": true, 00:12:06.498 "num_base_bdevs": 2, 00:12:06.498 "num_base_bdevs_discovered": 1, 00:12:06.498 "num_base_bdevs_operational": 1, 00:12:06.498 "base_bdevs_list": [ 00:12:06.498 { 00:12:06.498 "name": null, 00:12:06.498 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:06.498 "is_configured": false, 00:12:06.498 "data_offset": 0, 00:12:06.498 "data_size": 63488 00:12:06.498 }, 00:12:06.498 { 00:12:06.498 "name": "BaseBdev2", 00:12:06.498 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:12:06.498 "is_configured": true, 00:12:06.498 "data_offset": 2048, 00:12:06.498 "data_size": 63488 00:12:06.498 } 00:12:06.498 ] 00:12:06.498 }' 00:12:06.498 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:06.498 19:02:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:06.759 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:06.759 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:06.759 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:06.759 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:06.759 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:06.759 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:06.759 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:06.759 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:06.759 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:06.759 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:06.759 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:06.759 "name": "raid_bdev1", 00:12:06.759 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:12:06.759 "strip_size_kb": 0, 00:12:06.759 "state": "online", 00:12:06.759 "raid_level": "raid1", 00:12:06.759 "superblock": true, 00:12:06.759 "num_base_bdevs": 2, 00:12:06.759 "num_base_bdevs_discovered": 1, 00:12:06.759 "num_base_bdevs_operational": 1, 00:12:06.759 "base_bdevs_list": [ 00:12:06.759 { 00:12:06.759 "name": null, 00:12:06.759 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:06.759 "is_configured": false, 00:12:06.759 "data_offset": 0, 00:12:06.759 "data_size": 63488 00:12:06.759 }, 00:12:06.759 { 00:12:06.759 "name": "BaseBdev2", 00:12:06.759 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:12:06.759 "is_configured": true, 00:12:06.759 "data_offset": 2048, 00:12:06.759 "data_size": 63488 00:12:06.759 } 00:12:06.759 ] 00:12:06.759 }' 00:12:06.759 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:06.759 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:06.759 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:07.019 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:07.019 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:12:07.019 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.019 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:07.019 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.019 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:07.019 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.019 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:07.019 [2024-12-05 19:02:24.356787] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:07.019 [2024-12-05 19:02:24.356896] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:07.019 [2024-12-05 19:02:24.356924] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:12:07.019 [2024-12-05 19:02:24.356937] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:07.019 [2024-12-05 19:02:24.357333] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:07.019 [2024-12-05 19:02:24.357356] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:07.019 [2024-12-05 19:02:24.357430] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:12:07.019 [2024-12-05 19:02:24.357448] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:07.019 [2024-12-05 19:02:24.357465] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:07.019 [2024-12-05 19:02:24.357483] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:12:07.019 BaseBdev1 00:12:07.019 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.019 19:02:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:07.957 "name": "raid_bdev1", 00:12:07.957 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:12:07.957 "strip_size_kb": 0, 00:12:07.957 "state": "online", 00:12:07.957 "raid_level": "raid1", 00:12:07.957 "superblock": true, 00:12:07.957 "num_base_bdevs": 2, 00:12:07.957 "num_base_bdevs_discovered": 1, 00:12:07.957 "num_base_bdevs_operational": 1, 00:12:07.957 "base_bdevs_list": [ 00:12:07.957 { 00:12:07.957 "name": null, 00:12:07.957 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:07.957 "is_configured": false, 00:12:07.957 "data_offset": 0, 00:12:07.957 "data_size": 63488 00:12:07.957 }, 00:12:07.957 { 00:12:07.957 "name": "BaseBdev2", 00:12:07.957 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:12:07.957 "is_configured": true, 00:12:07.957 "data_offset": 2048, 00:12:07.957 "data_size": 63488 00:12:07.957 } 00:12:07.957 ] 00:12:07.957 }' 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:07.957 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:08.525 "name": "raid_bdev1", 00:12:08.525 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:12:08.525 "strip_size_kb": 0, 00:12:08.525 "state": "online", 00:12:08.525 "raid_level": "raid1", 00:12:08.525 "superblock": true, 00:12:08.525 "num_base_bdevs": 2, 00:12:08.525 "num_base_bdevs_discovered": 1, 00:12:08.525 "num_base_bdevs_operational": 1, 00:12:08.525 "base_bdevs_list": [ 00:12:08.525 { 00:12:08.525 "name": null, 00:12:08.525 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:08.525 "is_configured": false, 00:12:08.525 "data_offset": 0, 00:12:08.525 "data_size": 63488 00:12:08.525 }, 00:12:08.525 { 00:12:08.525 "name": "BaseBdev2", 00:12:08.525 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:12:08.525 "is_configured": true, 00:12:08.525 "data_offset": 2048, 00:12:08.525 "data_size": 63488 00:12:08.525 } 00:12:08.525 ] 00:12:08.525 }' 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # local es=0 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:08.525 [2024-12-05 19:02:25.990231] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:08.525 [2024-12-05 19:02:25.990392] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:08.525 [2024-12-05 19:02:25.990405] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:08.525 request: 00:12:08.525 { 00:12:08.525 "base_bdev": "BaseBdev1", 00:12:08.525 "raid_bdev": "raid_bdev1", 00:12:08.525 "method": "bdev_raid_add_base_bdev", 00:12:08.525 "req_id": 1 00:12:08.525 } 00:12:08.525 Got JSON-RPC error response 00:12:08.525 response: 00:12:08.525 { 00:12:08.525 "code": -22, 00:12:08.525 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:12:08.525 } 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # es=1 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:12:08.525 19:02:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:09.463 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:09.463 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:09.463 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:09.463 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:09.463 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:09.463 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:09.463 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:09.463 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:09.463 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:09.463 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:09.463 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.463 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.463 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.463 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.723 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.723 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:09.723 "name": "raid_bdev1", 00:12:09.723 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:12:09.723 "strip_size_kb": 0, 00:12:09.723 "state": "online", 00:12:09.723 "raid_level": "raid1", 00:12:09.723 "superblock": true, 00:12:09.723 "num_base_bdevs": 2, 00:12:09.723 "num_base_bdevs_discovered": 1, 00:12:09.723 "num_base_bdevs_operational": 1, 00:12:09.723 "base_bdevs_list": [ 00:12:09.723 { 00:12:09.723 "name": null, 00:12:09.723 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:09.723 "is_configured": false, 00:12:09.723 "data_offset": 0, 00:12:09.723 "data_size": 63488 00:12:09.723 }, 00:12:09.723 { 00:12:09.723 "name": "BaseBdev2", 00:12:09.723 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:12:09.723 "is_configured": true, 00:12:09.723 "data_offset": 2048, 00:12:09.723 "data_size": 63488 00:12:09.723 } 00:12:09.723 ] 00:12:09.723 }' 00:12:09.723 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:09.723 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.983 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:09.983 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:09.983 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:09.983 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:09.983 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:09.983 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.983 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:09.983 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.983 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.983 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:09.983 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:09.983 "name": "raid_bdev1", 00:12:09.983 "uuid": "491fdb9c-58b1-437f-8951-139f3391fd8e", 00:12:09.983 "strip_size_kb": 0, 00:12:09.983 "state": "online", 00:12:09.983 "raid_level": "raid1", 00:12:09.983 "superblock": true, 00:12:09.983 "num_base_bdevs": 2, 00:12:09.983 "num_base_bdevs_discovered": 1, 00:12:09.983 "num_base_bdevs_operational": 1, 00:12:09.983 "base_bdevs_list": [ 00:12:09.983 { 00:12:09.983 "name": null, 00:12:09.983 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:09.983 "is_configured": false, 00:12:09.983 "data_offset": 0, 00:12:09.983 "data_size": 63488 00:12:09.983 }, 00:12:09.983 { 00:12:09.983 "name": "BaseBdev2", 00:12:09.983 "uuid": "674d9443-5f83-58b3-91fe-bee30c1f0774", 00:12:09.983 "is_configured": true, 00:12:09.983 "data_offset": 2048, 00:12:09.983 "data_size": 63488 00:12:09.983 } 00:12:09.983 ] 00:12:09.983 }' 00:12:09.983 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:09.983 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:10.243 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:10.243 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:10.243 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 85989 00:12:10.243 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # '[' -z 85989 ']' 00:12:10.243 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # kill -0 85989 00:12:10.243 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # uname 00:12:10.243 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:10.243 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 85989 00:12:10.243 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:10.243 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:10.243 killing process with pid 85989 00:12:10.243 Received shutdown signal, test time was about 16.432330 seconds 00:12:10.243 00:12:10.243 Latency(us) 00:12:10.243 [2024-12-05T19:02:27.803Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:10.244 [2024-12-05T19:02:27.803Z] =================================================================================================================== 00:12:10.244 [2024-12-05T19:02:27.803Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:10.244 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 85989' 00:12:10.244 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@973 -- # kill 85989 00:12:10.244 [2024-12-05 19:02:27.626700] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:10.244 [2024-12-05 19:02:27.626830] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:10.244 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@978 -- # wait 85989 00:12:10.244 [2024-12-05 19:02:27.626891] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:10.244 [2024-12-05 19:02:27.626902] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:12:10.244 [2024-12-05 19:02:27.652850] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:12:10.504 00:12:10.504 real 0m18.245s 00:12:10.504 user 0m24.278s 00:12:10.504 sys 0m2.147s 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:10.504 ************************************ 00:12:10.504 END TEST raid_rebuild_test_sb_io 00:12:10.504 ************************************ 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.504 19:02:27 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:12:10.504 19:02:27 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 4 false false true 00:12:10.504 19:02:27 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:10.504 19:02:27 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:10.504 19:02:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:10.504 ************************************ 00:12:10.504 START TEST raid_rebuild_test 00:12:10.504 ************************************ 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 false false true 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=86552 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 86552 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 86552 ']' 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:10.504 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:10.504 19:02:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:10.504 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:10.504 Zero copy mechanism will not be used. 00:12:10.504 [2024-12-05 19:02:28.046185] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:12:10.504 [2024-12-05 19:02:28.046316] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86552 ] 00:12:10.765 [2024-12-05 19:02:28.203817] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:10.765 [2024-12-05 19:02:28.230223] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:10.765 [2024-12-05 19:02:28.274819] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:10.765 [2024-12-05 19:02:28.274858] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.336 BaseBdev1_malloc 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.336 [2024-12-05 19:02:28.868211] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:11.336 [2024-12-05 19:02:28.868369] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:11.336 [2024-12-05 19:02:28.868408] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:11.336 [2024-12-05 19:02:28.868422] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:11.336 [2024-12-05 19:02:28.870546] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:11.336 [2024-12-05 19:02:28.870589] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:11.336 BaseBdev1 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.336 BaseBdev2_malloc 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.336 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.336 [2024-12-05 19:02:28.893097] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:11.336 [2024-12-05 19:02:28.893157] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:11.336 [2024-12-05 19:02:28.893183] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:11.336 [2024-12-05 19:02:28.893194] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:11.597 [2024-12-05 19:02:28.895268] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:11.597 [2024-12-05 19:02:28.895381] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:11.597 BaseBdev2 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.597 BaseBdev3_malloc 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.597 [2024-12-05 19:02:28.921987] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:11.597 [2024-12-05 19:02:28.922104] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:11.597 [2024-12-05 19:02:28.922134] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:11.597 [2024-12-05 19:02:28.922145] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:11.597 [2024-12-05 19:02:28.924236] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:11.597 [2024-12-05 19:02:28.924277] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:11.597 BaseBdev3 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.597 BaseBdev4_malloc 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.597 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.597 [2024-12-05 19:02:28.965781] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:11.597 [2024-12-05 19:02:28.965883] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:11.597 [2024-12-05 19:02:28.965936] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:11.597 [2024-12-05 19:02:28.965963] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:11.598 [2024-12-05 19:02:28.969201] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:11.598 [2024-12-05 19:02:28.969337] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:11.598 BaseBdev4 00:12:11.598 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.598 19:02:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:11.598 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.598 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.598 spare_malloc 00:12:11.598 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.598 19:02:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:11.598 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.598 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.598 spare_delay 00:12:11.598 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.598 19:02:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:11.598 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.598 19:02:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.598 [2024-12-05 19:02:29.007324] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:11.598 [2024-12-05 19:02:29.007376] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:11.598 [2024-12-05 19:02:29.007397] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:12:11.598 [2024-12-05 19:02:29.007407] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:11.598 [2024-12-05 19:02:29.009513] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:11.598 [2024-12-05 19:02:29.009557] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:11.598 spare 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.598 [2024-12-05 19:02:29.019382] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:11.598 [2024-12-05 19:02:29.021253] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:11.598 [2024-12-05 19:02:29.021321] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:11.598 [2024-12-05 19:02:29.021371] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:11.598 [2024-12-05 19:02:29.021456] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:11.598 [2024-12-05 19:02:29.021471] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:11.598 [2024-12-05 19:02:29.021749] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:11.598 [2024-12-05 19:02:29.021907] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:11.598 [2024-12-05 19:02:29.021922] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:11.598 [2024-12-05 19:02:29.022053] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:11.598 "name": "raid_bdev1", 00:12:11.598 "uuid": "203b3618-5a7b-4a07-a7e3-702bc16241f0", 00:12:11.598 "strip_size_kb": 0, 00:12:11.598 "state": "online", 00:12:11.598 "raid_level": "raid1", 00:12:11.598 "superblock": false, 00:12:11.598 "num_base_bdevs": 4, 00:12:11.598 "num_base_bdevs_discovered": 4, 00:12:11.598 "num_base_bdevs_operational": 4, 00:12:11.598 "base_bdevs_list": [ 00:12:11.598 { 00:12:11.598 "name": "BaseBdev1", 00:12:11.598 "uuid": "bdee4bb6-0807-5e4a-bbc1-61fbcb95679e", 00:12:11.598 "is_configured": true, 00:12:11.598 "data_offset": 0, 00:12:11.598 "data_size": 65536 00:12:11.598 }, 00:12:11.598 { 00:12:11.598 "name": "BaseBdev2", 00:12:11.598 "uuid": "89149769-73bc-57cb-abcf-d6c512eeaafe", 00:12:11.598 "is_configured": true, 00:12:11.598 "data_offset": 0, 00:12:11.598 "data_size": 65536 00:12:11.598 }, 00:12:11.598 { 00:12:11.598 "name": "BaseBdev3", 00:12:11.598 "uuid": "87d565b7-d84c-53dd-b893-8efd1e02cdcb", 00:12:11.598 "is_configured": true, 00:12:11.598 "data_offset": 0, 00:12:11.598 "data_size": 65536 00:12:11.598 }, 00:12:11.598 { 00:12:11.598 "name": "BaseBdev4", 00:12:11.598 "uuid": "e0486e98-43b6-5df9-9836-27e8367987d2", 00:12:11.598 "is_configured": true, 00:12:11.598 "data_offset": 0, 00:12:11.598 "data_size": 65536 00:12:11.598 } 00:12:11.598 ] 00:12:11.598 }' 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:11.598 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:12.168 [2024-12-05 19:02:29.490861] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:12.168 19:02:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:12.427 [2024-12-05 19:02:29.730207] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:12:12.427 /dev/nbd0 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:12.427 1+0 records in 00:12:12.427 1+0 records out 00:12:12.427 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000262686 s, 15.6 MB/s 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:12.427 19:02:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:12:19.000 65536+0 records in 00:12:19.000 65536+0 records out 00:12:19.000 33554432 bytes (34 MB, 32 MiB) copied, 5.93404 s, 5.7 MB/s 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:19.000 [2024-12-05 19:02:35.922415] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.000 [2024-12-05 19:02:35.970383] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.000 19:02:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.000 19:02:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.000 19:02:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:19.000 "name": "raid_bdev1", 00:12:19.000 "uuid": "203b3618-5a7b-4a07-a7e3-702bc16241f0", 00:12:19.000 "strip_size_kb": 0, 00:12:19.000 "state": "online", 00:12:19.000 "raid_level": "raid1", 00:12:19.000 "superblock": false, 00:12:19.000 "num_base_bdevs": 4, 00:12:19.000 "num_base_bdevs_discovered": 3, 00:12:19.000 "num_base_bdevs_operational": 3, 00:12:19.000 "base_bdevs_list": [ 00:12:19.000 { 00:12:19.000 "name": null, 00:12:19.000 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:19.000 "is_configured": false, 00:12:19.000 "data_offset": 0, 00:12:19.000 "data_size": 65536 00:12:19.000 }, 00:12:19.000 { 00:12:19.000 "name": "BaseBdev2", 00:12:19.000 "uuid": "89149769-73bc-57cb-abcf-d6c512eeaafe", 00:12:19.000 "is_configured": true, 00:12:19.000 "data_offset": 0, 00:12:19.000 "data_size": 65536 00:12:19.000 }, 00:12:19.000 { 00:12:19.000 "name": "BaseBdev3", 00:12:19.000 "uuid": "87d565b7-d84c-53dd-b893-8efd1e02cdcb", 00:12:19.000 "is_configured": true, 00:12:19.000 "data_offset": 0, 00:12:19.000 "data_size": 65536 00:12:19.000 }, 00:12:19.000 { 00:12:19.000 "name": "BaseBdev4", 00:12:19.000 "uuid": "e0486e98-43b6-5df9-9836-27e8367987d2", 00:12:19.000 "is_configured": true, 00:12:19.000 "data_offset": 0, 00:12:19.000 "data_size": 65536 00:12:19.000 } 00:12:19.000 ] 00:12:19.000 }' 00:12:19.000 19:02:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:19.000 19:02:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.000 19:02:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:19.000 19:02:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.000 19:02:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.000 [2024-12-05 19:02:36.425717] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:19.000 [2024-12-05 19:02:36.430178] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d063c0 00:12:19.000 19:02:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.000 19:02:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:19.000 [2024-12-05 19:02:36.432197] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:19.941 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:19.941 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:19.941 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:19.941 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:19.941 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:19.941 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:19.941 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:19.941 19:02:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:19.941 19:02:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:19.941 19:02:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:19.941 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:19.941 "name": "raid_bdev1", 00:12:19.941 "uuid": "203b3618-5a7b-4a07-a7e3-702bc16241f0", 00:12:19.941 "strip_size_kb": 0, 00:12:19.941 "state": "online", 00:12:19.941 "raid_level": "raid1", 00:12:19.941 "superblock": false, 00:12:19.941 "num_base_bdevs": 4, 00:12:19.941 "num_base_bdevs_discovered": 4, 00:12:19.941 "num_base_bdevs_operational": 4, 00:12:19.941 "process": { 00:12:19.941 "type": "rebuild", 00:12:19.941 "target": "spare", 00:12:19.941 "progress": { 00:12:19.941 "blocks": 20480, 00:12:19.941 "percent": 31 00:12:19.941 } 00:12:19.941 }, 00:12:19.941 "base_bdevs_list": [ 00:12:19.941 { 00:12:19.941 "name": "spare", 00:12:19.941 "uuid": "3400e3e1-0fab-5d56-aa98-96cceaf70c2d", 00:12:19.941 "is_configured": true, 00:12:19.941 "data_offset": 0, 00:12:19.941 "data_size": 65536 00:12:19.941 }, 00:12:19.941 { 00:12:19.941 "name": "BaseBdev2", 00:12:19.941 "uuid": "89149769-73bc-57cb-abcf-d6c512eeaafe", 00:12:19.941 "is_configured": true, 00:12:19.941 "data_offset": 0, 00:12:19.941 "data_size": 65536 00:12:19.941 }, 00:12:19.941 { 00:12:19.941 "name": "BaseBdev3", 00:12:19.941 "uuid": "87d565b7-d84c-53dd-b893-8efd1e02cdcb", 00:12:19.941 "is_configured": true, 00:12:19.941 "data_offset": 0, 00:12:19.941 "data_size": 65536 00:12:19.941 }, 00:12:19.941 { 00:12:19.941 "name": "BaseBdev4", 00:12:19.941 "uuid": "e0486e98-43b6-5df9-9836-27e8367987d2", 00:12:19.941 "is_configured": true, 00:12:19.941 "data_offset": 0, 00:12:19.941 "data_size": 65536 00:12:19.941 } 00:12:19.941 ] 00:12:19.941 }' 00:12:19.941 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.201 [2024-12-05 19:02:37.592998] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:20.201 [2024-12-05 19:02:37.636913] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:20.201 [2024-12-05 19:02:37.636977] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:20.201 [2024-12-05 19:02:37.637014] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:20.201 [2024-12-05 19:02:37.637029] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.201 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:20.201 "name": "raid_bdev1", 00:12:20.201 "uuid": "203b3618-5a7b-4a07-a7e3-702bc16241f0", 00:12:20.201 "strip_size_kb": 0, 00:12:20.201 "state": "online", 00:12:20.201 "raid_level": "raid1", 00:12:20.201 "superblock": false, 00:12:20.201 "num_base_bdevs": 4, 00:12:20.202 "num_base_bdevs_discovered": 3, 00:12:20.202 "num_base_bdevs_operational": 3, 00:12:20.202 "base_bdevs_list": [ 00:12:20.202 { 00:12:20.202 "name": null, 00:12:20.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:20.202 "is_configured": false, 00:12:20.202 "data_offset": 0, 00:12:20.202 "data_size": 65536 00:12:20.202 }, 00:12:20.202 { 00:12:20.202 "name": "BaseBdev2", 00:12:20.202 "uuid": "89149769-73bc-57cb-abcf-d6c512eeaafe", 00:12:20.202 "is_configured": true, 00:12:20.202 "data_offset": 0, 00:12:20.202 "data_size": 65536 00:12:20.202 }, 00:12:20.202 { 00:12:20.202 "name": "BaseBdev3", 00:12:20.202 "uuid": "87d565b7-d84c-53dd-b893-8efd1e02cdcb", 00:12:20.202 "is_configured": true, 00:12:20.202 "data_offset": 0, 00:12:20.202 "data_size": 65536 00:12:20.202 }, 00:12:20.202 { 00:12:20.202 "name": "BaseBdev4", 00:12:20.202 "uuid": "e0486e98-43b6-5df9-9836-27e8367987d2", 00:12:20.202 "is_configured": true, 00:12:20.202 "data_offset": 0, 00:12:20.202 "data_size": 65536 00:12:20.202 } 00:12:20.202 ] 00:12:20.202 }' 00:12:20.202 19:02:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:20.202 19:02:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:20.771 "name": "raid_bdev1", 00:12:20.771 "uuid": "203b3618-5a7b-4a07-a7e3-702bc16241f0", 00:12:20.771 "strip_size_kb": 0, 00:12:20.771 "state": "online", 00:12:20.771 "raid_level": "raid1", 00:12:20.771 "superblock": false, 00:12:20.771 "num_base_bdevs": 4, 00:12:20.771 "num_base_bdevs_discovered": 3, 00:12:20.771 "num_base_bdevs_operational": 3, 00:12:20.771 "base_bdevs_list": [ 00:12:20.771 { 00:12:20.771 "name": null, 00:12:20.771 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:20.771 "is_configured": false, 00:12:20.771 "data_offset": 0, 00:12:20.771 "data_size": 65536 00:12:20.771 }, 00:12:20.771 { 00:12:20.771 "name": "BaseBdev2", 00:12:20.771 "uuid": "89149769-73bc-57cb-abcf-d6c512eeaafe", 00:12:20.771 "is_configured": true, 00:12:20.771 "data_offset": 0, 00:12:20.771 "data_size": 65536 00:12:20.771 }, 00:12:20.771 { 00:12:20.771 "name": "BaseBdev3", 00:12:20.771 "uuid": "87d565b7-d84c-53dd-b893-8efd1e02cdcb", 00:12:20.771 "is_configured": true, 00:12:20.771 "data_offset": 0, 00:12:20.771 "data_size": 65536 00:12:20.771 }, 00:12:20.771 { 00:12:20.771 "name": "BaseBdev4", 00:12:20.771 "uuid": "e0486e98-43b6-5df9-9836-27e8367987d2", 00:12:20.771 "is_configured": true, 00:12:20.771 "data_offset": 0, 00:12:20.771 "data_size": 65536 00:12:20.771 } 00:12:20.771 ] 00:12:20.771 }' 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:20.771 [2024-12-05 19:02:38.228497] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:20.771 [2024-12-05 19:02:38.232630] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d06490 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:20.771 19:02:38 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:20.771 [2024-12-05 19:02:38.234612] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:21.708 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:21.708 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:21.708 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:21.708 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:21.708 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:21.708 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:21.708 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:21.708 19:02:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:21.708 19:02:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:21.708 19:02:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:21.967 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:21.967 "name": "raid_bdev1", 00:12:21.967 "uuid": "203b3618-5a7b-4a07-a7e3-702bc16241f0", 00:12:21.967 "strip_size_kb": 0, 00:12:21.967 "state": "online", 00:12:21.967 "raid_level": "raid1", 00:12:21.968 "superblock": false, 00:12:21.968 "num_base_bdevs": 4, 00:12:21.968 "num_base_bdevs_discovered": 4, 00:12:21.968 "num_base_bdevs_operational": 4, 00:12:21.968 "process": { 00:12:21.968 "type": "rebuild", 00:12:21.968 "target": "spare", 00:12:21.968 "progress": { 00:12:21.968 "blocks": 20480, 00:12:21.968 "percent": 31 00:12:21.968 } 00:12:21.968 }, 00:12:21.968 "base_bdevs_list": [ 00:12:21.968 { 00:12:21.968 "name": "spare", 00:12:21.968 "uuid": "3400e3e1-0fab-5d56-aa98-96cceaf70c2d", 00:12:21.968 "is_configured": true, 00:12:21.968 "data_offset": 0, 00:12:21.968 "data_size": 65536 00:12:21.968 }, 00:12:21.968 { 00:12:21.968 "name": "BaseBdev2", 00:12:21.968 "uuid": "89149769-73bc-57cb-abcf-d6c512eeaafe", 00:12:21.968 "is_configured": true, 00:12:21.968 "data_offset": 0, 00:12:21.968 "data_size": 65536 00:12:21.968 }, 00:12:21.968 { 00:12:21.968 "name": "BaseBdev3", 00:12:21.968 "uuid": "87d565b7-d84c-53dd-b893-8efd1e02cdcb", 00:12:21.968 "is_configured": true, 00:12:21.968 "data_offset": 0, 00:12:21.968 "data_size": 65536 00:12:21.968 }, 00:12:21.968 { 00:12:21.968 "name": "BaseBdev4", 00:12:21.968 "uuid": "e0486e98-43b6-5df9-9836-27e8367987d2", 00:12:21.968 "is_configured": true, 00:12:21.968 "data_offset": 0, 00:12:21.968 "data_size": 65536 00:12:21.968 } 00:12:21.968 ] 00:12:21.968 }' 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:21.968 [2024-12-05 19:02:39.387063] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:21.968 [2024-12-05 19:02:39.438631] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000d06490 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:21.968 "name": "raid_bdev1", 00:12:21.968 "uuid": "203b3618-5a7b-4a07-a7e3-702bc16241f0", 00:12:21.968 "strip_size_kb": 0, 00:12:21.968 "state": "online", 00:12:21.968 "raid_level": "raid1", 00:12:21.968 "superblock": false, 00:12:21.968 "num_base_bdevs": 4, 00:12:21.968 "num_base_bdevs_discovered": 3, 00:12:21.968 "num_base_bdevs_operational": 3, 00:12:21.968 "process": { 00:12:21.968 "type": "rebuild", 00:12:21.968 "target": "spare", 00:12:21.968 "progress": { 00:12:21.968 "blocks": 24576, 00:12:21.968 "percent": 37 00:12:21.968 } 00:12:21.968 }, 00:12:21.968 "base_bdevs_list": [ 00:12:21.968 { 00:12:21.968 "name": "spare", 00:12:21.968 "uuid": "3400e3e1-0fab-5d56-aa98-96cceaf70c2d", 00:12:21.968 "is_configured": true, 00:12:21.968 "data_offset": 0, 00:12:21.968 "data_size": 65536 00:12:21.968 }, 00:12:21.968 { 00:12:21.968 "name": null, 00:12:21.968 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:21.968 "is_configured": false, 00:12:21.968 "data_offset": 0, 00:12:21.968 "data_size": 65536 00:12:21.968 }, 00:12:21.968 { 00:12:21.968 "name": "BaseBdev3", 00:12:21.968 "uuid": "87d565b7-d84c-53dd-b893-8efd1e02cdcb", 00:12:21.968 "is_configured": true, 00:12:21.968 "data_offset": 0, 00:12:21.968 "data_size": 65536 00:12:21.968 }, 00:12:21.968 { 00:12:21.968 "name": "BaseBdev4", 00:12:21.968 "uuid": "e0486e98-43b6-5df9-9836-27e8367987d2", 00:12:21.968 "is_configured": true, 00:12:21.968 "data_offset": 0, 00:12:21.968 "data_size": 65536 00:12:21.968 } 00:12:21.968 ] 00:12:21.968 }' 00:12:21.968 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=349 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:22.228 "name": "raid_bdev1", 00:12:22.228 "uuid": "203b3618-5a7b-4a07-a7e3-702bc16241f0", 00:12:22.228 "strip_size_kb": 0, 00:12:22.228 "state": "online", 00:12:22.228 "raid_level": "raid1", 00:12:22.228 "superblock": false, 00:12:22.228 "num_base_bdevs": 4, 00:12:22.228 "num_base_bdevs_discovered": 3, 00:12:22.228 "num_base_bdevs_operational": 3, 00:12:22.228 "process": { 00:12:22.228 "type": "rebuild", 00:12:22.228 "target": "spare", 00:12:22.228 "progress": { 00:12:22.228 "blocks": 26624, 00:12:22.228 "percent": 40 00:12:22.228 } 00:12:22.228 }, 00:12:22.228 "base_bdevs_list": [ 00:12:22.228 { 00:12:22.228 "name": "spare", 00:12:22.228 "uuid": "3400e3e1-0fab-5d56-aa98-96cceaf70c2d", 00:12:22.228 "is_configured": true, 00:12:22.228 "data_offset": 0, 00:12:22.228 "data_size": 65536 00:12:22.228 }, 00:12:22.228 { 00:12:22.228 "name": null, 00:12:22.228 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:22.228 "is_configured": false, 00:12:22.228 "data_offset": 0, 00:12:22.228 "data_size": 65536 00:12:22.228 }, 00:12:22.228 { 00:12:22.228 "name": "BaseBdev3", 00:12:22.228 "uuid": "87d565b7-d84c-53dd-b893-8efd1e02cdcb", 00:12:22.228 "is_configured": true, 00:12:22.228 "data_offset": 0, 00:12:22.228 "data_size": 65536 00:12:22.228 }, 00:12:22.228 { 00:12:22.228 "name": "BaseBdev4", 00:12:22.228 "uuid": "e0486e98-43b6-5df9-9836-27e8367987d2", 00:12:22.228 "is_configured": true, 00:12:22.228 "data_offset": 0, 00:12:22.228 "data_size": 65536 00:12:22.228 } 00:12:22.228 ] 00:12:22.228 }' 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:22.228 19:02:39 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:23.168 19:02:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:23.168 19:02:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:23.168 19:02:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:23.168 19:02:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:23.168 19:02:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:23.168 19:02:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:23.168 19:02:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:23.168 19:02:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:23.168 19:02:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:23.168 19:02:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:23.428 19:02:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:23.428 19:02:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:23.428 "name": "raid_bdev1", 00:12:23.428 "uuid": "203b3618-5a7b-4a07-a7e3-702bc16241f0", 00:12:23.428 "strip_size_kb": 0, 00:12:23.428 "state": "online", 00:12:23.428 "raid_level": "raid1", 00:12:23.428 "superblock": false, 00:12:23.428 "num_base_bdevs": 4, 00:12:23.428 "num_base_bdevs_discovered": 3, 00:12:23.428 "num_base_bdevs_operational": 3, 00:12:23.428 "process": { 00:12:23.428 "type": "rebuild", 00:12:23.428 "target": "spare", 00:12:23.428 "progress": { 00:12:23.428 "blocks": 49152, 00:12:23.428 "percent": 75 00:12:23.428 } 00:12:23.428 }, 00:12:23.428 "base_bdevs_list": [ 00:12:23.428 { 00:12:23.428 "name": "spare", 00:12:23.428 "uuid": "3400e3e1-0fab-5d56-aa98-96cceaf70c2d", 00:12:23.428 "is_configured": true, 00:12:23.428 "data_offset": 0, 00:12:23.428 "data_size": 65536 00:12:23.428 }, 00:12:23.428 { 00:12:23.428 "name": null, 00:12:23.428 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:23.428 "is_configured": false, 00:12:23.428 "data_offset": 0, 00:12:23.428 "data_size": 65536 00:12:23.428 }, 00:12:23.428 { 00:12:23.428 "name": "BaseBdev3", 00:12:23.428 "uuid": "87d565b7-d84c-53dd-b893-8efd1e02cdcb", 00:12:23.428 "is_configured": true, 00:12:23.428 "data_offset": 0, 00:12:23.428 "data_size": 65536 00:12:23.428 }, 00:12:23.428 { 00:12:23.428 "name": "BaseBdev4", 00:12:23.428 "uuid": "e0486e98-43b6-5df9-9836-27e8367987d2", 00:12:23.428 "is_configured": true, 00:12:23.428 "data_offset": 0, 00:12:23.428 "data_size": 65536 00:12:23.428 } 00:12:23.428 ] 00:12:23.428 }' 00:12:23.428 19:02:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:23.428 19:02:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:23.428 19:02:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:23.428 19:02:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:23.428 19:02:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:23.999 [2024-12-05 19:02:41.445764] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:23.999 [2024-12-05 19:02:41.445923] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:23.999 [2024-12-05 19:02:41.445998] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:24.569 "name": "raid_bdev1", 00:12:24.569 "uuid": "203b3618-5a7b-4a07-a7e3-702bc16241f0", 00:12:24.569 "strip_size_kb": 0, 00:12:24.569 "state": "online", 00:12:24.569 "raid_level": "raid1", 00:12:24.569 "superblock": false, 00:12:24.569 "num_base_bdevs": 4, 00:12:24.569 "num_base_bdevs_discovered": 3, 00:12:24.569 "num_base_bdevs_operational": 3, 00:12:24.569 "base_bdevs_list": [ 00:12:24.569 { 00:12:24.569 "name": "spare", 00:12:24.569 "uuid": "3400e3e1-0fab-5d56-aa98-96cceaf70c2d", 00:12:24.569 "is_configured": true, 00:12:24.569 "data_offset": 0, 00:12:24.569 "data_size": 65536 00:12:24.569 }, 00:12:24.569 { 00:12:24.569 "name": null, 00:12:24.569 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:24.569 "is_configured": false, 00:12:24.569 "data_offset": 0, 00:12:24.569 "data_size": 65536 00:12:24.569 }, 00:12:24.569 { 00:12:24.569 "name": "BaseBdev3", 00:12:24.569 "uuid": "87d565b7-d84c-53dd-b893-8efd1e02cdcb", 00:12:24.569 "is_configured": true, 00:12:24.569 "data_offset": 0, 00:12:24.569 "data_size": 65536 00:12:24.569 }, 00:12:24.569 { 00:12:24.569 "name": "BaseBdev4", 00:12:24.569 "uuid": "e0486e98-43b6-5df9-9836-27e8367987d2", 00:12:24.569 "is_configured": true, 00:12:24.569 "data_offset": 0, 00:12:24.569 "data_size": 65536 00:12:24.569 } 00:12:24.569 ] 00:12:24.569 }' 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:24.569 19:02:41 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:24.569 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:24.569 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:24.569 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:24.569 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:24.569 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:24.569 "name": "raid_bdev1", 00:12:24.569 "uuid": "203b3618-5a7b-4a07-a7e3-702bc16241f0", 00:12:24.569 "strip_size_kb": 0, 00:12:24.569 "state": "online", 00:12:24.569 "raid_level": "raid1", 00:12:24.569 "superblock": false, 00:12:24.569 "num_base_bdevs": 4, 00:12:24.569 "num_base_bdevs_discovered": 3, 00:12:24.569 "num_base_bdevs_operational": 3, 00:12:24.569 "base_bdevs_list": [ 00:12:24.569 { 00:12:24.569 "name": "spare", 00:12:24.569 "uuid": "3400e3e1-0fab-5d56-aa98-96cceaf70c2d", 00:12:24.569 "is_configured": true, 00:12:24.569 "data_offset": 0, 00:12:24.569 "data_size": 65536 00:12:24.569 }, 00:12:24.569 { 00:12:24.569 "name": null, 00:12:24.569 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:24.569 "is_configured": false, 00:12:24.569 "data_offset": 0, 00:12:24.569 "data_size": 65536 00:12:24.569 }, 00:12:24.569 { 00:12:24.569 "name": "BaseBdev3", 00:12:24.569 "uuid": "87d565b7-d84c-53dd-b893-8efd1e02cdcb", 00:12:24.569 "is_configured": true, 00:12:24.569 "data_offset": 0, 00:12:24.569 "data_size": 65536 00:12:24.569 }, 00:12:24.569 { 00:12:24.569 "name": "BaseBdev4", 00:12:24.569 "uuid": "e0486e98-43b6-5df9-9836-27e8367987d2", 00:12:24.569 "is_configured": true, 00:12:24.569 "data_offset": 0, 00:12:24.569 "data_size": 65536 00:12:24.569 } 00:12:24.569 ] 00:12:24.569 }' 00:12:24.569 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:24.569 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:24.569 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:24.829 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:24.829 "name": "raid_bdev1", 00:12:24.829 "uuid": "203b3618-5a7b-4a07-a7e3-702bc16241f0", 00:12:24.829 "strip_size_kb": 0, 00:12:24.829 "state": "online", 00:12:24.829 "raid_level": "raid1", 00:12:24.829 "superblock": false, 00:12:24.829 "num_base_bdevs": 4, 00:12:24.829 "num_base_bdevs_discovered": 3, 00:12:24.829 "num_base_bdevs_operational": 3, 00:12:24.829 "base_bdevs_list": [ 00:12:24.829 { 00:12:24.829 "name": "spare", 00:12:24.829 "uuid": "3400e3e1-0fab-5d56-aa98-96cceaf70c2d", 00:12:24.829 "is_configured": true, 00:12:24.829 "data_offset": 0, 00:12:24.829 "data_size": 65536 00:12:24.829 }, 00:12:24.829 { 00:12:24.829 "name": null, 00:12:24.829 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:24.829 "is_configured": false, 00:12:24.830 "data_offset": 0, 00:12:24.830 "data_size": 65536 00:12:24.830 }, 00:12:24.830 { 00:12:24.830 "name": "BaseBdev3", 00:12:24.830 "uuid": "87d565b7-d84c-53dd-b893-8efd1e02cdcb", 00:12:24.830 "is_configured": true, 00:12:24.830 "data_offset": 0, 00:12:24.830 "data_size": 65536 00:12:24.830 }, 00:12:24.830 { 00:12:24.830 "name": "BaseBdev4", 00:12:24.830 "uuid": "e0486e98-43b6-5df9-9836-27e8367987d2", 00:12:24.830 "is_configured": true, 00:12:24.830 "data_offset": 0, 00:12:24.830 "data_size": 65536 00:12:24.830 } 00:12:24.830 ] 00:12:24.830 }' 00:12:24.830 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:24.830 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:25.090 [2024-12-05 19:02:42.564129] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:25.090 [2024-12-05 19:02:42.564212] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:25.090 [2024-12-05 19:02:42.564332] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:25.090 [2024-12-05 19:02:42.564442] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:25.090 [2024-12-05 19:02:42.564504] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:25.090 19:02:42 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:25.350 /dev/nbd0 00:12:25.350 19:02:42 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:25.350 19:02:42 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:25.350 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:25.350 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:12:25.350 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:25.350 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:25.350 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:25.350 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:12:25.351 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:25.351 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:25.351 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:25.351 1+0 records in 00:12:25.351 1+0 records out 00:12:25.351 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.0005653 s, 7.2 MB/s 00:12:25.351 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:25.351 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:12:25.351 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:25.351 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:25.351 19:02:42 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:12:25.351 19:02:42 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:25.351 19:02:42 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:25.351 19:02:42 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:25.611 /dev/nbd1 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@877 -- # break 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:25.611 1+0 records in 00:12:25.611 1+0 records out 00:12:25.611 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000437319 s, 9.4 MB/s 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:25.611 19:02:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:25.871 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 86552 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 86552 ']' 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 86552 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 86552 00:12:26.131 killing process with pid 86552 00:12:26.131 Received shutdown signal, test time was about 60.000000 seconds 00:12:26.131 00:12:26.131 Latency(us) 00:12:26.131 [2024-12-05T19:02:43.690Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:26.131 [2024-12-05T19:02:43.690Z] =================================================================================================================== 00:12:26.131 [2024-12-05T19:02:43.690Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 86552' 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@973 -- # kill 86552 00:12:26.131 [2024-12-05 19:02:43.653118] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:26.131 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@978 -- # wait 86552 00:12:26.392 [2024-12-05 19:02:43.702948] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:26.392 19:02:43 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:12:26.392 00:12:26.392 real 0m15.974s 00:12:26.392 user 0m17.449s 00:12:26.392 sys 0m3.382s 00:12:26.392 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:26.392 ************************************ 00:12:26.392 END TEST raid_rebuild_test 00:12:26.392 ************************************ 00:12:26.392 19:02:43 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:26.653 19:02:43 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 4 true false true 00:12:26.653 19:02:43 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:26.653 19:02:43 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:26.653 19:02:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:26.653 ************************************ 00:12:26.653 START TEST raid_rebuild_test_sb 00:12:26.653 ************************************ 00:12:26.653 19:02:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 true false true 00:12:26.653 19:02:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:26.653 19:02:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:26.653 19:02:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:26.653 19:02:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:26.653 19:02:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:26.653 19:02:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=86890 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 86890 00:12:26.653 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 86890 ']' 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:26.653 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.653 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:26.653 Zero copy mechanism will not be used. 00:12:26.653 [2024-12-05 19:02:44.095603] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:12:26.654 [2024-12-05 19:02:44.095757] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86890 ] 00:12:26.913 [2024-12-05 19:02:44.252165] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:26.913 [2024-12-05 19:02:44.278758] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:26.913 [2024-12-05 19:02:44.323477] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:26.913 [2024-12-05 19:02:44.323517] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.484 BaseBdev1_malloc 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.484 [2024-12-05 19:02:44.925063] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:27.484 [2024-12-05 19:02:44.925185] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:27.484 [2024-12-05 19:02:44.925227] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:27.484 [2024-12-05 19:02:44.925240] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:27.484 [2024-12-05 19:02:44.927397] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:27.484 [2024-12-05 19:02:44.927438] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:27.484 BaseBdev1 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.484 BaseBdev2_malloc 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.484 [2024-12-05 19:02:44.953942] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:27.484 [2024-12-05 19:02:44.954077] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:27.484 [2024-12-05 19:02:44.954124] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:27.484 [2024-12-05 19:02:44.954161] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:27.484 [2024-12-05 19:02:44.956332] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:27.484 [2024-12-05 19:02:44.956437] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:27.484 BaseBdev2 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.484 BaseBdev3_malloc 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.484 [2024-12-05 19:02:44.982829] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:27.484 [2024-12-05 19:02:44.982946] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:27.484 [2024-12-05 19:02:44.982974] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:27.484 [2024-12-05 19:02:44.982985] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:27.484 [2024-12-05 19:02:44.985151] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:27.484 [2024-12-05 19:02:44.985190] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:27.484 BaseBdev3 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.484 19:02:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.484 BaseBdev4_malloc 00:12:27.484 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.484 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:27.484 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.484 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.484 [2024-12-05 19:02:45.031132] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:27.484 [2024-12-05 19:02:45.031233] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:27.485 [2024-12-05 19:02:45.031286] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:27.485 [2024-12-05 19:02:45.031312] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:27.485 [2024-12-05 19:02:45.035365] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:27.485 [2024-12-05 19:02:45.035422] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:27.485 BaseBdev4 00:12:27.485 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.485 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:27.485 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.485 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.743 spare_malloc 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.743 spare_delay 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.743 [2024-12-05 19:02:45.073588] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:27.743 [2024-12-05 19:02:45.073713] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:27.743 [2024-12-05 19:02:45.073740] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:12:27.743 [2024-12-05 19:02:45.073751] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:27.743 [2024-12-05 19:02:45.075901] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:27.743 [2024-12-05 19:02:45.075952] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:27.743 spare 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.743 [2024-12-05 19:02:45.085645] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:27.743 [2024-12-05 19:02:45.087572] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:27.743 [2024-12-05 19:02:45.087638] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:27.743 [2024-12-05 19:02:45.087722] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:27.743 [2024-12-05 19:02:45.087902] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:27.743 [2024-12-05 19:02:45.087919] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:27.743 [2024-12-05 19:02:45.088219] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:27.743 [2024-12-05 19:02:45.088390] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:27.743 [2024-12-05 19:02:45.088404] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:27.743 [2024-12-05 19:02:45.088541] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:27.743 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:27.744 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:27.744 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:27.744 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:27.744 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:27.744 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:27.744 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:27.744 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:27.744 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:27.744 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.744 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:27.744 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:27.744 "name": "raid_bdev1", 00:12:27.744 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:27.744 "strip_size_kb": 0, 00:12:27.744 "state": "online", 00:12:27.744 "raid_level": "raid1", 00:12:27.744 "superblock": true, 00:12:27.744 "num_base_bdevs": 4, 00:12:27.744 "num_base_bdevs_discovered": 4, 00:12:27.744 "num_base_bdevs_operational": 4, 00:12:27.744 "base_bdevs_list": [ 00:12:27.744 { 00:12:27.744 "name": "BaseBdev1", 00:12:27.744 "uuid": "2d14a9b6-1723-5cfd-9797-7d0af435164f", 00:12:27.744 "is_configured": true, 00:12:27.744 "data_offset": 2048, 00:12:27.744 "data_size": 63488 00:12:27.744 }, 00:12:27.744 { 00:12:27.744 "name": "BaseBdev2", 00:12:27.744 "uuid": "80e00738-beaf-5d43-b1fe-275e96eb02c6", 00:12:27.744 "is_configured": true, 00:12:27.744 "data_offset": 2048, 00:12:27.744 "data_size": 63488 00:12:27.744 }, 00:12:27.744 { 00:12:27.744 "name": "BaseBdev3", 00:12:27.744 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:27.744 "is_configured": true, 00:12:27.744 "data_offset": 2048, 00:12:27.744 "data_size": 63488 00:12:27.744 }, 00:12:27.744 { 00:12:27.744 "name": "BaseBdev4", 00:12:27.744 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:27.744 "is_configured": true, 00:12:27.744 "data_offset": 2048, 00:12:27.744 "data_size": 63488 00:12:27.744 } 00:12:27.744 ] 00:12:27.744 }' 00:12:27.744 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:27.744 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:28.312 [2024-12-05 19:02:45.569115] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:28.312 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:28.312 [2024-12-05 19:02:45.848384] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:12:28.312 /dev/nbd0 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:28.570 1+0 records in 00:12:28.570 1+0 records out 00:12:28.570 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000622773 s, 6.6 MB/s 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:28.570 19:02:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:12:35.229 63488+0 records in 00:12:35.229 63488+0 records out 00:12:35.229 32505856 bytes (33 MB, 31 MiB) copied, 5.68158 s, 5.7 MB/s 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:35.229 [2024-12-05 19:02:51.800168] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:35.229 [2024-12-05 19:02:51.836464] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:35.229 "name": "raid_bdev1", 00:12:35.229 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:35.229 "strip_size_kb": 0, 00:12:35.229 "state": "online", 00:12:35.229 "raid_level": "raid1", 00:12:35.229 "superblock": true, 00:12:35.229 "num_base_bdevs": 4, 00:12:35.229 "num_base_bdevs_discovered": 3, 00:12:35.229 "num_base_bdevs_operational": 3, 00:12:35.229 "base_bdevs_list": [ 00:12:35.229 { 00:12:35.229 "name": null, 00:12:35.229 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:35.229 "is_configured": false, 00:12:35.229 "data_offset": 0, 00:12:35.229 "data_size": 63488 00:12:35.229 }, 00:12:35.229 { 00:12:35.229 "name": "BaseBdev2", 00:12:35.229 "uuid": "80e00738-beaf-5d43-b1fe-275e96eb02c6", 00:12:35.229 "is_configured": true, 00:12:35.229 "data_offset": 2048, 00:12:35.229 "data_size": 63488 00:12:35.229 }, 00:12:35.229 { 00:12:35.229 "name": "BaseBdev3", 00:12:35.229 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:35.229 "is_configured": true, 00:12:35.229 "data_offset": 2048, 00:12:35.229 "data_size": 63488 00:12:35.229 }, 00:12:35.229 { 00:12:35.229 "name": "BaseBdev4", 00:12:35.229 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:35.229 "is_configured": true, 00:12:35.229 "data_offset": 2048, 00:12:35.229 "data_size": 63488 00:12:35.229 } 00:12:35.229 ] 00:12:35.229 }' 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:35.229 19:02:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:35.229 19:02:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:35.229 19:02:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.229 19:02:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:35.229 [2024-12-05 19:02:52.275753] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:35.229 [2024-12-05 19:02:52.280198] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e420 00:12:35.229 19:02:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.229 19:02:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:35.229 [2024-12-05 19:02:52.282267] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:35.799 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:35.799 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:35.799 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:35.799 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:35.799 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:35.799 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:35.799 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:35.799 19:02:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:35.799 19:02:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:35.799 19:02:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:35.799 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:35.799 "name": "raid_bdev1", 00:12:35.799 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:35.799 "strip_size_kb": 0, 00:12:35.799 "state": "online", 00:12:35.799 "raid_level": "raid1", 00:12:35.799 "superblock": true, 00:12:35.799 "num_base_bdevs": 4, 00:12:35.799 "num_base_bdevs_discovered": 4, 00:12:35.799 "num_base_bdevs_operational": 4, 00:12:35.799 "process": { 00:12:35.799 "type": "rebuild", 00:12:35.799 "target": "spare", 00:12:35.799 "progress": { 00:12:35.799 "blocks": 20480, 00:12:35.799 "percent": 32 00:12:35.799 } 00:12:35.799 }, 00:12:35.799 "base_bdevs_list": [ 00:12:35.799 { 00:12:35.799 "name": "spare", 00:12:35.799 "uuid": "f8699cd1-514b-5389-9f52-cb89540a8b0f", 00:12:35.799 "is_configured": true, 00:12:35.799 "data_offset": 2048, 00:12:35.799 "data_size": 63488 00:12:35.799 }, 00:12:35.800 { 00:12:35.800 "name": "BaseBdev2", 00:12:35.800 "uuid": "80e00738-beaf-5d43-b1fe-275e96eb02c6", 00:12:35.800 "is_configured": true, 00:12:35.800 "data_offset": 2048, 00:12:35.800 "data_size": 63488 00:12:35.800 }, 00:12:35.800 { 00:12:35.800 "name": "BaseBdev3", 00:12:35.800 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:35.800 "is_configured": true, 00:12:35.800 "data_offset": 2048, 00:12:35.800 "data_size": 63488 00:12:35.800 }, 00:12:35.800 { 00:12:35.800 "name": "BaseBdev4", 00:12:35.800 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:35.800 "is_configured": true, 00:12:35.800 "data_offset": 2048, 00:12:35.800 "data_size": 63488 00:12:35.800 } 00:12:35.800 ] 00:12:35.800 }' 00:12:35.800 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.060 [2024-12-05 19:02:53.419870] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:36.060 [2024-12-05 19:02:53.486938] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:36.060 [2024-12-05 19:02:53.487064] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:36.060 [2024-12-05 19:02:53.487114] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:36.060 [2024-12-05 19:02:53.487155] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:36.060 "name": "raid_bdev1", 00:12:36.060 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:36.060 "strip_size_kb": 0, 00:12:36.060 "state": "online", 00:12:36.060 "raid_level": "raid1", 00:12:36.060 "superblock": true, 00:12:36.060 "num_base_bdevs": 4, 00:12:36.060 "num_base_bdevs_discovered": 3, 00:12:36.060 "num_base_bdevs_operational": 3, 00:12:36.060 "base_bdevs_list": [ 00:12:36.060 { 00:12:36.060 "name": null, 00:12:36.060 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:36.060 "is_configured": false, 00:12:36.060 "data_offset": 0, 00:12:36.060 "data_size": 63488 00:12:36.060 }, 00:12:36.060 { 00:12:36.060 "name": "BaseBdev2", 00:12:36.060 "uuid": "80e00738-beaf-5d43-b1fe-275e96eb02c6", 00:12:36.060 "is_configured": true, 00:12:36.060 "data_offset": 2048, 00:12:36.060 "data_size": 63488 00:12:36.060 }, 00:12:36.060 { 00:12:36.060 "name": "BaseBdev3", 00:12:36.060 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:36.060 "is_configured": true, 00:12:36.060 "data_offset": 2048, 00:12:36.060 "data_size": 63488 00:12:36.060 }, 00:12:36.060 { 00:12:36.060 "name": "BaseBdev4", 00:12:36.060 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:36.060 "is_configured": true, 00:12:36.060 "data_offset": 2048, 00:12:36.060 "data_size": 63488 00:12:36.060 } 00:12:36.060 ] 00:12:36.060 }' 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:36.060 19:02:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.630 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:36.630 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:36.630 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:36.630 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:36.630 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:36.630 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:36.630 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:36.630 19:02:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.630 19:02:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.630 19:02:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.630 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:36.630 "name": "raid_bdev1", 00:12:36.630 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:36.630 "strip_size_kb": 0, 00:12:36.630 "state": "online", 00:12:36.630 "raid_level": "raid1", 00:12:36.630 "superblock": true, 00:12:36.630 "num_base_bdevs": 4, 00:12:36.630 "num_base_bdevs_discovered": 3, 00:12:36.630 "num_base_bdevs_operational": 3, 00:12:36.630 "base_bdevs_list": [ 00:12:36.630 { 00:12:36.630 "name": null, 00:12:36.630 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:36.630 "is_configured": false, 00:12:36.630 "data_offset": 0, 00:12:36.630 "data_size": 63488 00:12:36.630 }, 00:12:36.630 { 00:12:36.630 "name": "BaseBdev2", 00:12:36.630 "uuid": "80e00738-beaf-5d43-b1fe-275e96eb02c6", 00:12:36.630 "is_configured": true, 00:12:36.630 "data_offset": 2048, 00:12:36.630 "data_size": 63488 00:12:36.630 }, 00:12:36.630 { 00:12:36.630 "name": "BaseBdev3", 00:12:36.630 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:36.630 "is_configured": true, 00:12:36.630 "data_offset": 2048, 00:12:36.630 "data_size": 63488 00:12:36.630 }, 00:12:36.630 { 00:12:36.630 "name": "BaseBdev4", 00:12:36.630 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:36.630 "is_configured": true, 00:12:36.630 "data_offset": 2048, 00:12:36.630 "data_size": 63488 00:12:36.630 } 00:12:36.630 ] 00:12:36.630 }' 00:12:36.631 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:36.631 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:36.631 19:02:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:36.631 19:02:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:36.631 19:02:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:36.631 19:02:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:36.631 19:02:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:36.631 [2024-12-05 19:02:54.038559] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:36.631 [2024-12-05 19:02:54.042221] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e4f0 00:12:36.631 [2024-12-05 19:02:54.044168] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:36.631 19:02:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:36.631 19:02:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:37.571 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:37.571 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:37.571 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:37.571 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:37.571 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:37.571 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:37.571 19:02:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.571 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:37.571 19:02:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.571 19:02:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.571 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:37.571 "name": "raid_bdev1", 00:12:37.571 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:37.571 "strip_size_kb": 0, 00:12:37.571 "state": "online", 00:12:37.571 "raid_level": "raid1", 00:12:37.571 "superblock": true, 00:12:37.571 "num_base_bdevs": 4, 00:12:37.571 "num_base_bdevs_discovered": 4, 00:12:37.571 "num_base_bdevs_operational": 4, 00:12:37.571 "process": { 00:12:37.571 "type": "rebuild", 00:12:37.571 "target": "spare", 00:12:37.571 "progress": { 00:12:37.571 "blocks": 20480, 00:12:37.571 "percent": 32 00:12:37.571 } 00:12:37.571 }, 00:12:37.571 "base_bdevs_list": [ 00:12:37.571 { 00:12:37.571 "name": "spare", 00:12:37.571 "uuid": "f8699cd1-514b-5389-9f52-cb89540a8b0f", 00:12:37.571 "is_configured": true, 00:12:37.571 "data_offset": 2048, 00:12:37.571 "data_size": 63488 00:12:37.571 }, 00:12:37.571 { 00:12:37.571 "name": "BaseBdev2", 00:12:37.571 "uuid": "80e00738-beaf-5d43-b1fe-275e96eb02c6", 00:12:37.571 "is_configured": true, 00:12:37.571 "data_offset": 2048, 00:12:37.571 "data_size": 63488 00:12:37.571 }, 00:12:37.571 { 00:12:37.571 "name": "BaseBdev3", 00:12:37.571 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:37.571 "is_configured": true, 00:12:37.571 "data_offset": 2048, 00:12:37.571 "data_size": 63488 00:12:37.571 }, 00:12:37.571 { 00:12:37.571 "name": "BaseBdev4", 00:12:37.571 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:37.571 "is_configured": true, 00:12:37.571 "data_offset": 2048, 00:12:37.571 "data_size": 63488 00:12:37.571 } 00:12:37.571 ] 00:12:37.571 }' 00:12:37.571 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:12:37.831 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.831 [2024-12-05 19:02:55.215378] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:37.831 [2024-12-05 19:02:55.348154] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000c3e4f0 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:37.831 19:02:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:38.091 "name": "raid_bdev1", 00:12:38.091 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:38.091 "strip_size_kb": 0, 00:12:38.091 "state": "online", 00:12:38.091 "raid_level": "raid1", 00:12:38.091 "superblock": true, 00:12:38.091 "num_base_bdevs": 4, 00:12:38.091 "num_base_bdevs_discovered": 3, 00:12:38.091 "num_base_bdevs_operational": 3, 00:12:38.091 "process": { 00:12:38.091 "type": "rebuild", 00:12:38.091 "target": "spare", 00:12:38.091 "progress": { 00:12:38.091 "blocks": 24576, 00:12:38.091 "percent": 38 00:12:38.091 } 00:12:38.091 }, 00:12:38.091 "base_bdevs_list": [ 00:12:38.091 { 00:12:38.091 "name": "spare", 00:12:38.091 "uuid": "f8699cd1-514b-5389-9f52-cb89540a8b0f", 00:12:38.091 "is_configured": true, 00:12:38.091 "data_offset": 2048, 00:12:38.091 "data_size": 63488 00:12:38.091 }, 00:12:38.091 { 00:12:38.091 "name": null, 00:12:38.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:38.091 "is_configured": false, 00:12:38.091 "data_offset": 0, 00:12:38.091 "data_size": 63488 00:12:38.091 }, 00:12:38.091 { 00:12:38.091 "name": "BaseBdev3", 00:12:38.091 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:38.091 "is_configured": true, 00:12:38.091 "data_offset": 2048, 00:12:38.091 "data_size": 63488 00:12:38.091 }, 00:12:38.091 { 00:12:38.091 "name": "BaseBdev4", 00:12:38.091 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:38.091 "is_configured": true, 00:12:38.091 "data_offset": 2048, 00:12:38.091 "data_size": 63488 00:12:38.091 } 00:12:38.091 ] 00:12:38.091 }' 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=365 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:38.091 "name": "raid_bdev1", 00:12:38.091 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:38.091 "strip_size_kb": 0, 00:12:38.091 "state": "online", 00:12:38.091 "raid_level": "raid1", 00:12:38.091 "superblock": true, 00:12:38.091 "num_base_bdevs": 4, 00:12:38.091 "num_base_bdevs_discovered": 3, 00:12:38.091 "num_base_bdevs_operational": 3, 00:12:38.091 "process": { 00:12:38.091 "type": "rebuild", 00:12:38.091 "target": "spare", 00:12:38.091 "progress": { 00:12:38.091 "blocks": 26624, 00:12:38.091 "percent": 41 00:12:38.091 } 00:12:38.091 }, 00:12:38.091 "base_bdevs_list": [ 00:12:38.091 { 00:12:38.091 "name": "spare", 00:12:38.091 "uuid": "f8699cd1-514b-5389-9f52-cb89540a8b0f", 00:12:38.091 "is_configured": true, 00:12:38.091 "data_offset": 2048, 00:12:38.091 "data_size": 63488 00:12:38.091 }, 00:12:38.091 { 00:12:38.091 "name": null, 00:12:38.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:38.091 "is_configured": false, 00:12:38.091 "data_offset": 0, 00:12:38.091 "data_size": 63488 00:12:38.091 }, 00:12:38.091 { 00:12:38.091 "name": "BaseBdev3", 00:12:38.091 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:38.091 "is_configured": true, 00:12:38.091 "data_offset": 2048, 00:12:38.091 "data_size": 63488 00:12:38.091 }, 00:12:38.091 { 00:12:38.091 "name": "BaseBdev4", 00:12:38.091 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:38.091 "is_configured": true, 00:12:38.091 "data_offset": 2048, 00:12:38.091 "data_size": 63488 00:12:38.091 } 00:12:38.091 ] 00:12:38.091 }' 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:38.091 19:02:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:39.469 19:02:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:39.469 19:02:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:39.469 19:02:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:39.469 19:02:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:39.469 19:02:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:39.469 19:02:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:39.469 19:02:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:39.469 19:02:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:39.469 19:02:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:39.469 19:02:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:39.469 19:02:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.469 19:02:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:39.469 "name": "raid_bdev1", 00:12:39.469 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:39.469 "strip_size_kb": 0, 00:12:39.469 "state": "online", 00:12:39.469 "raid_level": "raid1", 00:12:39.469 "superblock": true, 00:12:39.469 "num_base_bdevs": 4, 00:12:39.469 "num_base_bdevs_discovered": 3, 00:12:39.469 "num_base_bdevs_operational": 3, 00:12:39.469 "process": { 00:12:39.469 "type": "rebuild", 00:12:39.469 "target": "spare", 00:12:39.469 "progress": { 00:12:39.469 "blocks": 49152, 00:12:39.469 "percent": 77 00:12:39.469 } 00:12:39.469 }, 00:12:39.469 "base_bdevs_list": [ 00:12:39.469 { 00:12:39.469 "name": "spare", 00:12:39.469 "uuid": "f8699cd1-514b-5389-9f52-cb89540a8b0f", 00:12:39.469 "is_configured": true, 00:12:39.469 "data_offset": 2048, 00:12:39.469 "data_size": 63488 00:12:39.469 }, 00:12:39.469 { 00:12:39.469 "name": null, 00:12:39.469 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:39.469 "is_configured": false, 00:12:39.469 "data_offset": 0, 00:12:39.469 "data_size": 63488 00:12:39.469 }, 00:12:39.469 { 00:12:39.469 "name": "BaseBdev3", 00:12:39.470 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:39.470 "is_configured": true, 00:12:39.470 "data_offset": 2048, 00:12:39.470 "data_size": 63488 00:12:39.470 }, 00:12:39.470 { 00:12:39.470 "name": "BaseBdev4", 00:12:39.470 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:39.470 "is_configured": true, 00:12:39.470 "data_offset": 2048, 00:12:39.470 "data_size": 63488 00:12:39.470 } 00:12:39.470 ] 00:12:39.470 }' 00:12:39.470 19:02:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:39.470 19:02:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:39.470 19:02:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:39.470 19:02:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:39.470 19:02:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:39.728 [2024-12-05 19:02:57.254916] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:39.728 [2024-12-05 19:02:57.254991] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:39.728 [2024-12-05 19:02:57.255097] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:40.296 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:40.296 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:40.296 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:40.296 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:40.296 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:40.296 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:40.296 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.296 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:40.296 19:02:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:40.296 19:02:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:40.296 19:02:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:40.296 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:40.296 "name": "raid_bdev1", 00:12:40.296 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:40.296 "strip_size_kb": 0, 00:12:40.296 "state": "online", 00:12:40.296 "raid_level": "raid1", 00:12:40.296 "superblock": true, 00:12:40.296 "num_base_bdevs": 4, 00:12:40.296 "num_base_bdevs_discovered": 3, 00:12:40.296 "num_base_bdevs_operational": 3, 00:12:40.296 "base_bdevs_list": [ 00:12:40.296 { 00:12:40.296 "name": "spare", 00:12:40.296 "uuid": "f8699cd1-514b-5389-9f52-cb89540a8b0f", 00:12:40.296 "is_configured": true, 00:12:40.296 "data_offset": 2048, 00:12:40.296 "data_size": 63488 00:12:40.296 }, 00:12:40.296 { 00:12:40.296 "name": null, 00:12:40.296 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:40.296 "is_configured": false, 00:12:40.296 "data_offset": 0, 00:12:40.296 "data_size": 63488 00:12:40.296 }, 00:12:40.296 { 00:12:40.296 "name": "BaseBdev3", 00:12:40.296 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:40.296 "is_configured": true, 00:12:40.296 "data_offset": 2048, 00:12:40.296 "data_size": 63488 00:12:40.296 }, 00:12:40.296 { 00:12:40.296 "name": "BaseBdev4", 00:12:40.296 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:40.296 "is_configured": true, 00:12:40.296 "data_offset": 2048, 00:12:40.296 "data_size": 63488 00:12:40.296 } 00:12:40.296 ] 00:12:40.296 }' 00:12:40.296 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:40.296 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:40.296 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:40.556 "name": "raid_bdev1", 00:12:40.556 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:40.556 "strip_size_kb": 0, 00:12:40.556 "state": "online", 00:12:40.556 "raid_level": "raid1", 00:12:40.556 "superblock": true, 00:12:40.556 "num_base_bdevs": 4, 00:12:40.556 "num_base_bdevs_discovered": 3, 00:12:40.556 "num_base_bdevs_operational": 3, 00:12:40.556 "base_bdevs_list": [ 00:12:40.556 { 00:12:40.556 "name": "spare", 00:12:40.556 "uuid": "f8699cd1-514b-5389-9f52-cb89540a8b0f", 00:12:40.556 "is_configured": true, 00:12:40.556 "data_offset": 2048, 00:12:40.556 "data_size": 63488 00:12:40.556 }, 00:12:40.556 { 00:12:40.556 "name": null, 00:12:40.556 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:40.556 "is_configured": false, 00:12:40.556 "data_offset": 0, 00:12:40.556 "data_size": 63488 00:12:40.556 }, 00:12:40.556 { 00:12:40.556 "name": "BaseBdev3", 00:12:40.556 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:40.556 "is_configured": true, 00:12:40.556 "data_offset": 2048, 00:12:40.556 "data_size": 63488 00:12:40.556 }, 00:12:40.556 { 00:12:40.556 "name": "BaseBdev4", 00:12:40.556 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:40.556 "is_configured": true, 00:12:40.556 "data_offset": 2048, 00:12:40.556 "data_size": 63488 00:12:40.556 } 00:12:40.556 ] 00:12:40.556 }' 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:40.556 19:02:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:40.556 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:40.556 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:40.556 "name": "raid_bdev1", 00:12:40.556 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:40.556 "strip_size_kb": 0, 00:12:40.556 "state": "online", 00:12:40.556 "raid_level": "raid1", 00:12:40.556 "superblock": true, 00:12:40.556 "num_base_bdevs": 4, 00:12:40.556 "num_base_bdevs_discovered": 3, 00:12:40.556 "num_base_bdevs_operational": 3, 00:12:40.556 "base_bdevs_list": [ 00:12:40.556 { 00:12:40.556 "name": "spare", 00:12:40.556 "uuid": "f8699cd1-514b-5389-9f52-cb89540a8b0f", 00:12:40.556 "is_configured": true, 00:12:40.556 "data_offset": 2048, 00:12:40.556 "data_size": 63488 00:12:40.556 }, 00:12:40.556 { 00:12:40.556 "name": null, 00:12:40.556 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:40.556 "is_configured": false, 00:12:40.556 "data_offset": 0, 00:12:40.556 "data_size": 63488 00:12:40.556 }, 00:12:40.556 { 00:12:40.556 "name": "BaseBdev3", 00:12:40.556 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:40.556 "is_configured": true, 00:12:40.556 "data_offset": 2048, 00:12:40.556 "data_size": 63488 00:12:40.556 }, 00:12:40.556 { 00:12:40.556 "name": "BaseBdev4", 00:12:40.556 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:40.556 "is_configured": true, 00:12:40.556 "data_offset": 2048, 00:12:40.556 "data_size": 63488 00:12:40.556 } 00:12:40.556 ] 00:12:40.556 }' 00:12:40.556 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:40.556 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:41.122 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:41.122 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:41.122 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:41.122 [2024-12-05 19:02:58.444952] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:41.122 [2024-12-05 19:02:58.445032] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:41.122 [2024-12-05 19:02:58.445143] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:41.122 [2024-12-05 19:02:58.445252] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:41.122 [2024-12-05 19:02:58.445318] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:41.122 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:41.122 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:41.122 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:12:41.122 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:41.122 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:41.122 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:41.122 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:41.123 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:41.123 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:41.123 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:41.123 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:41.123 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:41.123 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:41.123 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:41.123 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:41.123 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:41.123 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:41.123 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:41.123 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:41.382 /dev/nbd0 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:41.382 1+0 records in 00:12:41.382 1+0 records out 00:12:41.382 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000552308 s, 7.4 MB/s 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:41.382 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:41.641 /dev/nbd1 00:12:41.641 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:41.641 19:02:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:41.641 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:12:41.641 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:12:41.641 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:12:41.641 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:12:41.641 19:02:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:41.641 1+0 records in 00:12:41.641 1+0 records out 00:12:41.641 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000518934 s, 7.9 MB/s 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:41.641 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:41.900 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:41.900 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:41.900 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:41.900 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:41.900 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:41.900 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:41.900 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:41.900 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:41.900 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:41.900 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:42.160 [2024-12-05 19:02:59.530857] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:42.160 [2024-12-05 19:02:59.530944] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:42.160 [2024-12-05 19:02:59.530965] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:12:42.160 [2024-12-05 19:02:59.530980] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:42.160 [2024-12-05 19:02:59.533252] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:42.160 [2024-12-05 19:02:59.533300] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:42.160 [2024-12-05 19:02:59.533399] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:42.160 [2024-12-05 19:02:59.533446] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:42.160 [2024-12-05 19:02:59.533564] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:42.160 [2024-12-05 19:02:59.533672] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:42.160 spare 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:42.160 [2024-12-05 19:02:59.633572] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:12:42.160 [2024-12-05 19:02:59.633605] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:42.160 [2024-12-05 19:02:59.633900] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeb00 00:12:42.160 [2024-12-05 19:02:59.634061] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:12:42.160 [2024-12-05 19:02:59.634073] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:12:42.160 [2024-12-05 19:02:59.634212] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.160 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:42.160 "name": "raid_bdev1", 00:12:42.160 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:42.160 "strip_size_kb": 0, 00:12:42.160 "state": "online", 00:12:42.160 "raid_level": "raid1", 00:12:42.160 "superblock": true, 00:12:42.160 "num_base_bdevs": 4, 00:12:42.160 "num_base_bdevs_discovered": 3, 00:12:42.160 "num_base_bdevs_operational": 3, 00:12:42.160 "base_bdevs_list": [ 00:12:42.160 { 00:12:42.160 "name": "spare", 00:12:42.160 "uuid": "f8699cd1-514b-5389-9f52-cb89540a8b0f", 00:12:42.160 "is_configured": true, 00:12:42.160 "data_offset": 2048, 00:12:42.160 "data_size": 63488 00:12:42.160 }, 00:12:42.160 { 00:12:42.160 "name": null, 00:12:42.160 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:42.160 "is_configured": false, 00:12:42.160 "data_offset": 2048, 00:12:42.160 "data_size": 63488 00:12:42.160 }, 00:12:42.160 { 00:12:42.160 "name": "BaseBdev3", 00:12:42.160 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:42.160 "is_configured": true, 00:12:42.160 "data_offset": 2048, 00:12:42.160 "data_size": 63488 00:12:42.160 }, 00:12:42.160 { 00:12:42.160 "name": "BaseBdev4", 00:12:42.160 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:42.160 "is_configured": true, 00:12:42.161 "data_offset": 2048, 00:12:42.161 "data_size": 63488 00:12:42.161 } 00:12:42.161 ] 00:12:42.161 }' 00:12:42.161 19:02:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:42.161 19:02:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:42.730 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:42.730 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:42.730 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:42.730 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:42.730 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:42.730 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:42.730 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:42.730 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.730 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:42.730 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.730 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:42.730 "name": "raid_bdev1", 00:12:42.730 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:42.730 "strip_size_kb": 0, 00:12:42.730 "state": "online", 00:12:42.730 "raid_level": "raid1", 00:12:42.730 "superblock": true, 00:12:42.730 "num_base_bdevs": 4, 00:12:42.730 "num_base_bdevs_discovered": 3, 00:12:42.730 "num_base_bdevs_operational": 3, 00:12:42.730 "base_bdevs_list": [ 00:12:42.730 { 00:12:42.730 "name": "spare", 00:12:42.730 "uuid": "f8699cd1-514b-5389-9f52-cb89540a8b0f", 00:12:42.730 "is_configured": true, 00:12:42.730 "data_offset": 2048, 00:12:42.730 "data_size": 63488 00:12:42.730 }, 00:12:42.730 { 00:12:42.730 "name": null, 00:12:42.730 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:42.730 "is_configured": false, 00:12:42.730 "data_offset": 2048, 00:12:42.730 "data_size": 63488 00:12:42.731 }, 00:12:42.731 { 00:12:42.731 "name": "BaseBdev3", 00:12:42.731 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:42.731 "is_configured": true, 00:12:42.731 "data_offset": 2048, 00:12:42.731 "data_size": 63488 00:12:42.731 }, 00:12:42.731 { 00:12:42.731 "name": "BaseBdev4", 00:12:42.731 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:42.731 "is_configured": true, 00:12:42.731 "data_offset": 2048, 00:12:42.731 "data_size": 63488 00:12:42.731 } 00:12:42.731 ] 00:12:42.731 }' 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:42.731 [2024-12-05 19:03:00.257681] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:42.731 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:42.991 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:42.991 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:42.991 "name": "raid_bdev1", 00:12:42.991 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:42.991 "strip_size_kb": 0, 00:12:42.991 "state": "online", 00:12:42.991 "raid_level": "raid1", 00:12:42.991 "superblock": true, 00:12:42.991 "num_base_bdevs": 4, 00:12:42.991 "num_base_bdevs_discovered": 2, 00:12:42.991 "num_base_bdevs_operational": 2, 00:12:42.991 "base_bdevs_list": [ 00:12:42.991 { 00:12:42.991 "name": null, 00:12:42.991 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:42.991 "is_configured": false, 00:12:42.991 "data_offset": 0, 00:12:42.991 "data_size": 63488 00:12:42.991 }, 00:12:42.991 { 00:12:42.991 "name": null, 00:12:42.991 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:42.991 "is_configured": false, 00:12:42.991 "data_offset": 2048, 00:12:42.991 "data_size": 63488 00:12:42.991 }, 00:12:42.991 { 00:12:42.991 "name": "BaseBdev3", 00:12:42.991 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:42.991 "is_configured": true, 00:12:42.991 "data_offset": 2048, 00:12:42.991 "data_size": 63488 00:12:42.991 }, 00:12:42.991 { 00:12:42.991 "name": "BaseBdev4", 00:12:42.991 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:42.991 "is_configured": true, 00:12:42.991 "data_offset": 2048, 00:12:42.991 "data_size": 63488 00:12:42.991 } 00:12:42.991 ] 00:12:42.991 }' 00:12:42.991 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:42.991 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:43.251 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:43.251 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:43.251 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:43.251 [2024-12-05 19:03:00.692917] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:43.251 [2024-12-05 19:03:00.693068] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:12:43.251 [2024-12-05 19:03:00.693091] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:43.251 [2024-12-05 19:03:00.693128] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:43.251 [2024-12-05 19:03:00.697341] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caebd0 00:12:43.251 19:03:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:43.251 19:03:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:43.251 [2024-12-05 19:03:00.699095] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:44.191 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:44.191 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:44.191 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:44.191 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:44.191 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:44.191 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.191 19:03:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.191 19:03:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.191 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:44.191 19:03:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:44.452 "name": "raid_bdev1", 00:12:44.452 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:44.452 "strip_size_kb": 0, 00:12:44.452 "state": "online", 00:12:44.452 "raid_level": "raid1", 00:12:44.452 "superblock": true, 00:12:44.452 "num_base_bdevs": 4, 00:12:44.452 "num_base_bdevs_discovered": 3, 00:12:44.452 "num_base_bdevs_operational": 3, 00:12:44.452 "process": { 00:12:44.452 "type": "rebuild", 00:12:44.452 "target": "spare", 00:12:44.452 "progress": { 00:12:44.452 "blocks": 20480, 00:12:44.452 "percent": 32 00:12:44.452 } 00:12:44.452 }, 00:12:44.452 "base_bdevs_list": [ 00:12:44.452 { 00:12:44.452 "name": "spare", 00:12:44.452 "uuid": "f8699cd1-514b-5389-9f52-cb89540a8b0f", 00:12:44.452 "is_configured": true, 00:12:44.452 "data_offset": 2048, 00:12:44.452 "data_size": 63488 00:12:44.452 }, 00:12:44.452 { 00:12:44.452 "name": null, 00:12:44.452 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:44.452 "is_configured": false, 00:12:44.452 "data_offset": 2048, 00:12:44.452 "data_size": 63488 00:12:44.452 }, 00:12:44.452 { 00:12:44.452 "name": "BaseBdev3", 00:12:44.452 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:44.452 "is_configured": true, 00:12:44.452 "data_offset": 2048, 00:12:44.452 "data_size": 63488 00:12:44.452 }, 00:12:44.452 { 00:12:44.452 "name": "BaseBdev4", 00:12:44.452 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:44.452 "is_configured": true, 00:12:44.452 "data_offset": 2048, 00:12:44.452 "data_size": 63488 00:12:44.452 } 00:12:44.452 ] 00:12:44.452 }' 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.452 [2024-12-05 19:03:01.860132] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:44.452 [2024-12-05 19:03:01.903187] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:44.452 [2024-12-05 19:03:01.903251] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:44.452 [2024-12-05 19:03:01.903268] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:44.452 [2024-12-05 19:03:01.903280] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:44.452 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:44.453 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:44.453 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:44.453 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:44.453 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.453 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:44.453 19:03:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:44.453 19:03:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:44.453 19:03:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:44.453 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:44.453 "name": "raid_bdev1", 00:12:44.453 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:44.453 "strip_size_kb": 0, 00:12:44.453 "state": "online", 00:12:44.453 "raid_level": "raid1", 00:12:44.453 "superblock": true, 00:12:44.453 "num_base_bdevs": 4, 00:12:44.453 "num_base_bdevs_discovered": 2, 00:12:44.453 "num_base_bdevs_operational": 2, 00:12:44.453 "base_bdevs_list": [ 00:12:44.453 { 00:12:44.453 "name": null, 00:12:44.453 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:44.453 "is_configured": false, 00:12:44.453 "data_offset": 0, 00:12:44.453 "data_size": 63488 00:12:44.453 }, 00:12:44.453 { 00:12:44.453 "name": null, 00:12:44.453 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:44.453 "is_configured": false, 00:12:44.453 "data_offset": 2048, 00:12:44.453 "data_size": 63488 00:12:44.453 }, 00:12:44.453 { 00:12:44.453 "name": "BaseBdev3", 00:12:44.453 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:44.453 "is_configured": true, 00:12:44.453 "data_offset": 2048, 00:12:44.453 "data_size": 63488 00:12:44.453 }, 00:12:44.453 { 00:12:44.453 "name": "BaseBdev4", 00:12:44.453 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:44.453 "is_configured": true, 00:12:44.453 "data_offset": 2048, 00:12:44.453 "data_size": 63488 00:12:44.453 } 00:12:44.453 ] 00:12:44.453 }' 00:12:44.453 19:03:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:44.453 19:03:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.023 19:03:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:45.023 19:03:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.023 19:03:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.023 [2024-12-05 19:03:02.362500] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:45.023 [2024-12-05 19:03:02.362633] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:45.023 [2024-12-05 19:03:02.362689] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:12:45.023 [2024-12-05 19:03:02.362746] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:45.023 [2024-12-05 19:03:02.363227] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:45.023 [2024-12-05 19:03:02.363298] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:45.023 [2024-12-05 19:03:02.363421] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:45.023 [2024-12-05 19:03:02.363475] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:12:45.023 [2024-12-05 19:03:02.363525] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:45.023 [2024-12-05 19:03:02.363580] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:45.023 [2024-12-05 19:03:02.367609] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeca0 00:12:45.023 spare 00:12:45.023 19:03:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.023 19:03:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:12:45.023 [2024-12-05 19:03:02.369610] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:45.963 "name": "raid_bdev1", 00:12:45.963 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:45.963 "strip_size_kb": 0, 00:12:45.963 "state": "online", 00:12:45.963 "raid_level": "raid1", 00:12:45.963 "superblock": true, 00:12:45.963 "num_base_bdevs": 4, 00:12:45.963 "num_base_bdevs_discovered": 3, 00:12:45.963 "num_base_bdevs_operational": 3, 00:12:45.963 "process": { 00:12:45.963 "type": "rebuild", 00:12:45.963 "target": "spare", 00:12:45.963 "progress": { 00:12:45.963 "blocks": 20480, 00:12:45.963 "percent": 32 00:12:45.963 } 00:12:45.963 }, 00:12:45.963 "base_bdevs_list": [ 00:12:45.963 { 00:12:45.963 "name": "spare", 00:12:45.963 "uuid": "f8699cd1-514b-5389-9f52-cb89540a8b0f", 00:12:45.963 "is_configured": true, 00:12:45.963 "data_offset": 2048, 00:12:45.963 "data_size": 63488 00:12:45.963 }, 00:12:45.963 { 00:12:45.963 "name": null, 00:12:45.963 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:45.963 "is_configured": false, 00:12:45.963 "data_offset": 2048, 00:12:45.963 "data_size": 63488 00:12:45.963 }, 00:12:45.963 { 00:12:45.963 "name": "BaseBdev3", 00:12:45.963 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:45.963 "is_configured": true, 00:12:45.963 "data_offset": 2048, 00:12:45.963 "data_size": 63488 00:12:45.963 }, 00:12:45.963 { 00:12:45.963 "name": "BaseBdev4", 00:12:45.963 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:45.963 "is_configured": true, 00:12:45.963 "data_offset": 2048, 00:12:45.963 "data_size": 63488 00:12:45.963 } 00:12:45.963 ] 00:12:45.963 }' 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:45.963 19:03:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.223 [2024-12-05 19:03:03.522266] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:46.223 [2024-12-05 19:03:03.573724] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:46.223 [2024-12-05 19:03:03.573855] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:46.223 [2024-12-05 19:03:03.573902] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:46.223 [2024-12-05 19:03:03.573928] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.223 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:46.223 "name": "raid_bdev1", 00:12:46.223 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:46.223 "strip_size_kb": 0, 00:12:46.223 "state": "online", 00:12:46.223 "raid_level": "raid1", 00:12:46.223 "superblock": true, 00:12:46.223 "num_base_bdevs": 4, 00:12:46.223 "num_base_bdevs_discovered": 2, 00:12:46.223 "num_base_bdevs_operational": 2, 00:12:46.223 "base_bdevs_list": [ 00:12:46.223 { 00:12:46.223 "name": null, 00:12:46.223 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:46.223 "is_configured": false, 00:12:46.223 "data_offset": 0, 00:12:46.223 "data_size": 63488 00:12:46.223 }, 00:12:46.223 { 00:12:46.223 "name": null, 00:12:46.223 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:46.223 "is_configured": false, 00:12:46.223 "data_offset": 2048, 00:12:46.224 "data_size": 63488 00:12:46.224 }, 00:12:46.224 { 00:12:46.224 "name": "BaseBdev3", 00:12:46.224 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:46.224 "is_configured": true, 00:12:46.224 "data_offset": 2048, 00:12:46.224 "data_size": 63488 00:12:46.224 }, 00:12:46.224 { 00:12:46.224 "name": "BaseBdev4", 00:12:46.224 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:46.224 "is_configured": true, 00:12:46.224 "data_offset": 2048, 00:12:46.224 "data_size": 63488 00:12:46.224 } 00:12:46.224 ] 00:12:46.224 }' 00:12:46.224 19:03:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:46.224 19:03:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.794 19:03:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:46.794 19:03:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:46.794 19:03:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:46.794 19:03:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:46.794 19:03:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:46.794 19:03:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:46.794 19:03:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:46.794 19:03:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.794 19:03:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.794 19:03:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.794 19:03:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:46.794 "name": "raid_bdev1", 00:12:46.794 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:46.794 "strip_size_kb": 0, 00:12:46.794 "state": "online", 00:12:46.794 "raid_level": "raid1", 00:12:46.794 "superblock": true, 00:12:46.794 "num_base_bdevs": 4, 00:12:46.794 "num_base_bdevs_discovered": 2, 00:12:46.794 "num_base_bdevs_operational": 2, 00:12:46.794 "base_bdevs_list": [ 00:12:46.794 { 00:12:46.794 "name": null, 00:12:46.794 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:46.794 "is_configured": false, 00:12:46.794 "data_offset": 0, 00:12:46.794 "data_size": 63488 00:12:46.794 }, 00:12:46.794 { 00:12:46.794 "name": null, 00:12:46.794 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:46.794 "is_configured": false, 00:12:46.794 "data_offset": 2048, 00:12:46.794 "data_size": 63488 00:12:46.794 }, 00:12:46.794 { 00:12:46.794 "name": "BaseBdev3", 00:12:46.794 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:46.794 "is_configured": true, 00:12:46.794 "data_offset": 2048, 00:12:46.794 "data_size": 63488 00:12:46.794 }, 00:12:46.794 { 00:12:46.794 "name": "BaseBdev4", 00:12:46.794 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:46.794 "is_configured": true, 00:12:46.794 "data_offset": 2048, 00:12:46.794 "data_size": 63488 00:12:46.794 } 00:12:46.794 ] 00:12:46.794 }' 00:12:46.794 19:03:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:46.794 19:03:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:46.795 19:03:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:46.795 19:03:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:46.795 19:03:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:12:46.795 19:03:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.795 19:03:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.795 19:03:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.795 19:03:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:46.795 19:03:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:46.795 19:03:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:46.795 [2024-12-05 19:03:04.224909] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:46.795 [2024-12-05 19:03:04.225034] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:46.795 [2024-12-05 19:03:04.225063] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:12:46.795 [2024-12-05 19:03:04.225074] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:46.795 [2024-12-05 19:03:04.225494] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:46.795 [2024-12-05 19:03:04.225515] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:46.795 [2024-12-05 19:03:04.225592] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:12:46.795 [2024-12-05 19:03:04.225606] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:12:46.795 [2024-12-05 19:03:04.225616] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:46.795 [2024-12-05 19:03:04.225627] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:12:46.795 BaseBdev1 00:12:46.795 19:03:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:46.795 19:03:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:47.736 "name": "raid_bdev1", 00:12:47.736 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:47.736 "strip_size_kb": 0, 00:12:47.736 "state": "online", 00:12:47.736 "raid_level": "raid1", 00:12:47.736 "superblock": true, 00:12:47.736 "num_base_bdevs": 4, 00:12:47.736 "num_base_bdevs_discovered": 2, 00:12:47.736 "num_base_bdevs_operational": 2, 00:12:47.736 "base_bdevs_list": [ 00:12:47.736 { 00:12:47.736 "name": null, 00:12:47.736 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:47.736 "is_configured": false, 00:12:47.736 "data_offset": 0, 00:12:47.736 "data_size": 63488 00:12:47.736 }, 00:12:47.736 { 00:12:47.736 "name": null, 00:12:47.736 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:47.736 "is_configured": false, 00:12:47.736 "data_offset": 2048, 00:12:47.736 "data_size": 63488 00:12:47.736 }, 00:12:47.736 { 00:12:47.736 "name": "BaseBdev3", 00:12:47.736 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:47.736 "is_configured": true, 00:12:47.736 "data_offset": 2048, 00:12:47.736 "data_size": 63488 00:12:47.736 }, 00:12:47.736 { 00:12:47.736 "name": "BaseBdev4", 00:12:47.736 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:47.736 "is_configured": true, 00:12:47.736 "data_offset": 2048, 00:12:47.736 "data_size": 63488 00:12:47.736 } 00:12:47.736 ] 00:12:47.736 }' 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:47.736 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:48.301 "name": "raid_bdev1", 00:12:48.301 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:48.301 "strip_size_kb": 0, 00:12:48.301 "state": "online", 00:12:48.301 "raid_level": "raid1", 00:12:48.301 "superblock": true, 00:12:48.301 "num_base_bdevs": 4, 00:12:48.301 "num_base_bdevs_discovered": 2, 00:12:48.301 "num_base_bdevs_operational": 2, 00:12:48.301 "base_bdevs_list": [ 00:12:48.301 { 00:12:48.301 "name": null, 00:12:48.301 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:48.301 "is_configured": false, 00:12:48.301 "data_offset": 0, 00:12:48.301 "data_size": 63488 00:12:48.301 }, 00:12:48.301 { 00:12:48.301 "name": null, 00:12:48.301 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:48.301 "is_configured": false, 00:12:48.301 "data_offset": 2048, 00:12:48.301 "data_size": 63488 00:12:48.301 }, 00:12:48.301 { 00:12:48.301 "name": "BaseBdev3", 00:12:48.301 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:48.301 "is_configured": true, 00:12:48.301 "data_offset": 2048, 00:12:48.301 "data_size": 63488 00:12:48.301 }, 00:12:48.301 { 00:12:48.301 "name": "BaseBdev4", 00:12:48.301 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:48.301 "is_configured": true, 00:12:48.301 "data_offset": 2048, 00:12:48.301 "data_size": 63488 00:12:48.301 } 00:12:48.301 ] 00:12:48.301 }' 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:48.301 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:12:48.302 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:48.302 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:12:48.302 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:48.302 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:12:48.302 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:48.302 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:48.302 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:48.302 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:48.302 [2024-12-05 19:03:05.818319] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:48.302 [2024-12-05 19:03:05.818474] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:12:48.302 [2024-12-05 19:03:05.818493] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:48.302 request: 00:12:48.302 { 00:12:48.302 "base_bdev": "BaseBdev1", 00:12:48.302 "raid_bdev": "raid_bdev1", 00:12:48.302 "method": "bdev_raid_add_base_bdev", 00:12:48.302 "req_id": 1 00:12:48.302 } 00:12:48.302 Got JSON-RPC error response 00:12:48.302 response: 00:12:48.302 { 00:12:48.302 "code": -22, 00:12:48.302 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:12:48.302 } 00:12:48.302 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:12:48.302 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:12:48.302 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:12:48.302 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:12:48.302 19:03:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:12:48.302 19:03:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:49.701 19:03:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:49.701 19:03:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:49.701 19:03:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:49.701 19:03:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:49.701 19:03:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:49.701 19:03:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:49.702 19:03:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:49.702 19:03:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:49.702 19:03:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:49.702 19:03:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:49.702 19:03:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.702 19:03:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.702 19:03:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.702 19:03:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:49.702 19:03:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.702 19:03:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:49.702 "name": "raid_bdev1", 00:12:49.702 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:49.702 "strip_size_kb": 0, 00:12:49.702 "state": "online", 00:12:49.702 "raid_level": "raid1", 00:12:49.702 "superblock": true, 00:12:49.702 "num_base_bdevs": 4, 00:12:49.702 "num_base_bdevs_discovered": 2, 00:12:49.702 "num_base_bdevs_operational": 2, 00:12:49.702 "base_bdevs_list": [ 00:12:49.702 { 00:12:49.702 "name": null, 00:12:49.702 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:49.702 "is_configured": false, 00:12:49.702 "data_offset": 0, 00:12:49.702 "data_size": 63488 00:12:49.702 }, 00:12:49.702 { 00:12:49.702 "name": null, 00:12:49.702 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:49.702 "is_configured": false, 00:12:49.702 "data_offset": 2048, 00:12:49.702 "data_size": 63488 00:12:49.702 }, 00:12:49.702 { 00:12:49.702 "name": "BaseBdev3", 00:12:49.702 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:49.702 "is_configured": true, 00:12:49.702 "data_offset": 2048, 00:12:49.702 "data_size": 63488 00:12:49.702 }, 00:12:49.702 { 00:12:49.702 "name": "BaseBdev4", 00:12:49.702 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:49.702 "is_configured": true, 00:12:49.702 "data_offset": 2048, 00:12:49.702 "data_size": 63488 00:12:49.702 } 00:12:49.702 ] 00:12:49.702 }' 00:12:49.702 19:03:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:49.702 19:03:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.961 19:03:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:49.961 19:03:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:49.961 19:03:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:49.961 19:03:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:49.961 19:03:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:49.961 19:03:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:49.962 19:03:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.962 19:03:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:49.962 19:03:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:49.962 19:03:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:49.962 19:03:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:49.962 "name": "raid_bdev1", 00:12:49.962 "uuid": "cd4b2291-a4a1-481e-b7b1-567a4dab4fbb", 00:12:49.962 "strip_size_kb": 0, 00:12:49.962 "state": "online", 00:12:49.962 "raid_level": "raid1", 00:12:49.962 "superblock": true, 00:12:49.962 "num_base_bdevs": 4, 00:12:49.962 "num_base_bdevs_discovered": 2, 00:12:49.962 "num_base_bdevs_operational": 2, 00:12:49.962 "base_bdevs_list": [ 00:12:49.962 { 00:12:49.962 "name": null, 00:12:49.962 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:49.962 "is_configured": false, 00:12:49.962 "data_offset": 0, 00:12:49.962 "data_size": 63488 00:12:49.962 }, 00:12:49.962 { 00:12:49.962 "name": null, 00:12:49.962 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:49.962 "is_configured": false, 00:12:49.962 "data_offset": 2048, 00:12:49.962 "data_size": 63488 00:12:49.962 }, 00:12:49.962 { 00:12:49.962 "name": "BaseBdev3", 00:12:49.962 "uuid": "6662bdd5-a881-514b-a51c-c0b75bcf74fa", 00:12:49.962 "is_configured": true, 00:12:49.962 "data_offset": 2048, 00:12:49.962 "data_size": 63488 00:12:49.962 }, 00:12:49.962 { 00:12:49.962 "name": "BaseBdev4", 00:12:49.962 "uuid": "19de7d3d-24d9-5447-b101-7a57cb32019d", 00:12:49.962 "is_configured": true, 00:12:49.962 "data_offset": 2048, 00:12:49.962 "data_size": 63488 00:12:49.962 } 00:12:49.962 ] 00:12:49.962 }' 00:12:49.962 19:03:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:49.962 19:03:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:49.962 19:03:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:49.962 19:03:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:49.962 19:03:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 86890 00:12:49.962 19:03:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 86890 ']' 00:12:49.962 19:03:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 86890 00:12:49.962 19:03:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:12:49.962 19:03:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:49.962 19:03:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 86890 00:12:50.222 19:03:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:50.222 19:03:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:50.222 19:03:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 86890' 00:12:50.222 killing process with pid 86890 00:12:50.222 19:03:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 86890 00:12:50.222 Received shutdown signal, test time was about 60.000000 seconds 00:12:50.222 00:12:50.222 Latency(us) 00:12:50.222 [2024-12-05T19:03:07.781Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:50.222 [2024-12-05T19:03:07.781Z] =================================================================================================================== 00:12:50.222 [2024-12-05T19:03:07.781Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:50.222 [2024-12-05 19:03:07.523492] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:50.222 [2024-12-05 19:03:07.523634] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:50.222 19:03:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 86890 00:12:50.222 [2024-12-05 19:03:07.523749] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:50.222 [2024-12-05 19:03:07.523764] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:12:50.222 [2024-12-05 19:03:07.574189] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:12:50.483 00:12:50.483 real 0m23.792s 00:12:50.483 user 0m28.696s 00:12:50.483 sys 0m4.128s 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:50.483 ************************************ 00:12:50.483 END TEST raid_rebuild_test_sb 00:12:50.483 ************************************ 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:50.483 19:03:07 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 4 false true true 00:12:50.483 19:03:07 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:12:50.483 19:03:07 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:50.483 19:03:07 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:50.483 ************************************ 00:12:50.483 START TEST raid_rebuild_test_io 00:12:50.483 ************************************ 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 false true true 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=87494 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 87494 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # '[' -z 87494 ']' 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:50.483 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:50.483 19:03:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:50.483 [2024-12-05 19:03:07.967202] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:12:50.483 [2024-12-05 19:03:07.967462] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87494 ] 00:12:50.483 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:50.483 Zero copy mechanism will not be used. 00:12:50.744 [2024-12-05 19:03:08.121956] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:50.744 [2024-12-05 19:03:08.147626] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:50.744 [2024-12-05 19:03:08.192334] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:50.744 [2024-12-05 19:03:08.192373] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # return 0 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.314 BaseBdev1_malloc 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.314 [2024-12-05 19:03:08.805700] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:51.314 [2024-12-05 19:03:08.805761] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:51.314 [2024-12-05 19:03:08.805795] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:51.314 [2024-12-05 19:03:08.805821] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:51.314 [2024-12-05 19:03:08.807925] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:51.314 [2024-12-05 19:03:08.807980] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:51.314 BaseBdev1 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.314 BaseBdev2_malloc 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.314 [2024-12-05 19:03:08.834442] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:51.314 [2024-12-05 19:03:08.834501] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:51.314 [2024-12-05 19:03:08.834525] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:51.314 [2024-12-05 19:03:08.834535] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:51.314 [2024-12-05 19:03:08.836676] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:51.314 [2024-12-05 19:03:08.836739] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:51.314 BaseBdev2 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.314 BaseBdev3_malloc 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.314 [2024-12-05 19:03:08.863136] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:51.314 [2024-12-05 19:03:08.863193] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:51.314 [2024-12-05 19:03:08.863217] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:51.314 [2024-12-05 19:03:08.863228] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:51.314 [2024-12-05 19:03:08.865279] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:51.314 [2024-12-05 19:03:08.865318] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:51.314 BaseBdev3 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.314 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.575 BaseBdev4_malloc 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.575 [2024-12-05 19:03:08.911077] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:51.575 [2024-12-05 19:03:08.911181] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:51.575 [2024-12-05 19:03:08.911233] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:51.575 [2024-12-05 19:03:08.911259] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:51.575 [2024-12-05 19:03:08.915094] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:51.575 [2024-12-05 19:03:08.915149] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:51.575 BaseBdev4 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.575 spare_malloc 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.575 spare_delay 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.575 [2024-12-05 19:03:08.953222] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:51.575 [2024-12-05 19:03:08.953322] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:51.575 [2024-12-05 19:03:08.953348] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:12:51.575 [2024-12-05 19:03:08.953358] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:51.575 [2024-12-05 19:03:08.955473] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:51.575 [2024-12-05 19:03:08.955513] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:51.575 spare 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.575 [2024-12-05 19:03:08.965269] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:51.575 [2024-12-05 19:03:08.967079] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:51.575 [2024-12-05 19:03:08.967152] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:51.575 [2024-12-05 19:03:08.967207] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:51.575 [2024-12-05 19:03:08.967295] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:51.575 [2024-12-05 19:03:08.967304] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:51.575 [2024-12-05 19:03:08.967583] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:51.575 [2024-12-05 19:03:08.967725] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:51.575 [2024-12-05 19:03:08.967739] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:51.575 [2024-12-05 19:03:08.967884] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.575 19:03:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:51.575 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:51.575 "name": "raid_bdev1", 00:12:51.575 "uuid": "41715d1d-e1e1-4ceb-b212-31d1d4785dab", 00:12:51.575 "strip_size_kb": 0, 00:12:51.575 "state": "online", 00:12:51.575 "raid_level": "raid1", 00:12:51.575 "superblock": false, 00:12:51.575 "num_base_bdevs": 4, 00:12:51.575 "num_base_bdevs_discovered": 4, 00:12:51.575 "num_base_bdevs_operational": 4, 00:12:51.575 "base_bdevs_list": [ 00:12:51.575 { 00:12:51.575 "name": "BaseBdev1", 00:12:51.575 "uuid": "23a17e81-cee6-5a2a-b1db-6405d8bfe478", 00:12:51.575 "is_configured": true, 00:12:51.575 "data_offset": 0, 00:12:51.575 "data_size": 65536 00:12:51.575 }, 00:12:51.575 { 00:12:51.575 "name": "BaseBdev2", 00:12:51.575 "uuid": "a6875699-e811-5fd7-b3ef-04ccf6c8d0c1", 00:12:51.575 "is_configured": true, 00:12:51.575 "data_offset": 0, 00:12:51.575 "data_size": 65536 00:12:51.575 }, 00:12:51.575 { 00:12:51.575 "name": "BaseBdev3", 00:12:51.575 "uuid": "4559b304-b699-5e2c-8133-42629be61bbf", 00:12:51.575 "is_configured": true, 00:12:51.575 "data_offset": 0, 00:12:51.575 "data_size": 65536 00:12:51.575 }, 00:12:51.575 { 00:12:51.575 "name": "BaseBdev4", 00:12:51.575 "uuid": "b61c0311-8c43-5e48-82d5-ce72db132f66", 00:12:51.575 "is_configured": true, 00:12:51.575 "data_offset": 0, 00:12:51.575 "data_size": 65536 00:12:51.575 } 00:12:51.575 ] 00:12:51.575 }' 00:12:51.575 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:51.575 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.145 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:52.145 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:52.145 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.145 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.145 [2024-12-05 19:03:09.428794] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:52.145 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.145 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:12:52.145 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.146 [2024-12-05 19:03:09.528305] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:52.146 "name": "raid_bdev1", 00:12:52.146 "uuid": "41715d1d-e1e1-4ceb-b212-31d1d4785dab", 00:12:52.146 "strip_size_kb": 0, 00:12:52.146 "state": "online", 00:12:52.146 "raid_level": "raid1", 00:12:52.146 "superblock": false, 00:12:52.146 "num_base_bdevs": 4, 00:12:52.146 "num_base_bdevs_discovered": 3, 00:12:52.146 "num_base_bdevs_operational": 3, 00:12:52.146 "base_bdevs_list": [ 00:12:52.146 { 00:12:52.146 "name": null, 00:12:52.146 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:52.146 "is_configured": false, 00:12:52.146 "data_offset": 0, 00:12:52.146 "data_size": 65536 00:12:52.146 }, 00:12:52.146 { 00:12:52.146 "name": "BaseBdev2", 00:12:52.146 "uuid": "a6875699-e811-5fd7-b3ef-04ccf6c8d0c1", 00:12:52.146 "is_configured": true, 00:12:52.146 "data_offset": 0, 00:12:52.146 "data_size": 65536 00:12:52.146 }, 00:12:52.146 { 00:12:52.146 "name": "BaseBdev3", 00:12:52.146 "uuid": "4559b304-b699-5e2c-8133-42629be61bbf", 00:12:52.146 "is_configured": true, 00:12:52.146 "data_offset": 0, 00:12:52.146 "data_size": 65536 00:12:52.146 }, 00:12:52.146 { 00:12:52.146 "name": "BaseBdev4", 00:12:52.146 "uuid": "b61c0311-8c43-5e48-82d5-ce72db132f66", 00:12:52.146 "is_configured": true, 00:12:52.146 "data_offset": 0, 00:12:52.146 "data_size": 65536 00:12:52.146 } 00:12:52.146 ] 00:12:52.146 }' 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:52.146 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.146 [2024-12-05 19:03:09.602229] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:12:52.146 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:52.146 Zero copy mechanism will not be used. 00:12:52.146 Running I/O for 60 seconds... 00:12:52.714 19:03:09 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:52.714 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:52.714 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.714 [2024-12-05 19:03:09.971716] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:52.714 19:03:09 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:52.714 19:03:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:52.714 [2024-12-05 19:03:10.013963] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002940 00:12:52.714 [2024-12-05 19:03:10.016179] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:52.715 [2024-12-05 19:03:10.130967] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:52.715 [2024-12-05 19:03:10.131462] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:52.974 [2024-12-05 19:03:10.362136] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:53.494 169.00 IOPS, 507.00 MiB/s [2024-12-05T19:03:11.053Z] [2024-12-05 19:03:10.837634] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:53.494 [2024-12-05 19:03:10.838237] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:53.494 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:53.494 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:53.494 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:53.494 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:53.494 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:53.494 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.494 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.494 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.494 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.494 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:53.754 "name": "raid_bdev1", 00:12:53.754 "uuid": "41715d1d-e1e1-4ceb-b212-31d1d4785dab", 00:12:53.754 "strip_size_kb": 0, 00:12:53.754 "state": "online", 00:12:53.754 "raid_level": "raid1", 00:12:53.754 "superblock": false, 00:12:53.754 "num_base_bdevs": 4, 00:12:53.754 "num_base_bdevs_discovered": 4, 00:12:53.754 "num_base_bdevs_operational": 4, 00:12:53.754 "process": { 00:12:53.754 "type": "rebuild", 00:12:53.754 "target": "spare", 00:12:53.754 "progress": { 00:12:53.754 "blocks": 10240, 00:12:53.754 "percent": 15 00:12:53.754 } 00:12:53.754 }, 00:12:53.754 "base_bdevs_list": [ 00:12:53.754 { 00:12:53.754 "name": "spare", 00:12:53.754 "uuid": "dab5acb0-6f3c-5bc5-86e7-432eb51dcf7d", 00:12:53.754 "is_configured": true, 00:12:53.754 "data_offset": 0, 00:12:53.754 "data_size": 65536 00:12:53.754 }, 00:12:53.754 { 00:12:53.754 "name": "BaseBdev2", 00:12:53.754 "uuid": "a6875699-e811-5fd7-b3ef-04ccf6c8d0c1", 00:12:53.754 "is_configured": true, 00:12:53.754 "data_offset": 0, 00:12:53.754 "data_size": 65536 00:12:53.754 }, 00:12:53.754 { 00:12:53.754 "name": "BaseBdev3", 00:12:53.754 "uuid": "4559b304-b699-5e2c-8133-42629be61bbf", 00:12:53.754 "is_configured": true, 00:12:53.754 "data_offset": 0, 00:12:53.754 "data_size": 65536 00:12:53.754 }, 00:12:53.754 { 00:12:53.754 "name": "BaseBdev4", 00:12:53.754 "uuid": "b61c0311-8c43-5e48-82d5-ce72db132f66", 00:12:53.754 "is_configured": true, 00:12:53.754 "data_offset": 0, 00:12:53.754 "data_size": 65536 00:12:53.754 } 00:12:53.754 ] 00:12:53.754 }' 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.754 [2024-12-05 19:03:11.149707] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:53.754 [2024-12-05 19:03:11.150878] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:53.754 [2024-12-05 19:03:11.157993] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:53.754 [2024-12-05 19:03:11.159907] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:53.754 [2024-12-05 19:03:11.159968] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:53.754 [2024-12-05 19:03:11.159981] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:53.754 [2024-12-05 19:03:11.183570] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002870 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:53.754 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.755 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.755 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:53.755 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:53.755 "name": "raid_bdev1", 00:12:53.755 "uuid": "41715d1d-e1e1-4ceb-b212-31d1d4785dab", 00:12:53.755 "strip_size_kb": 0, 00:12:53.755 "state": "online", 00:12:53.755 "raid_level": "raid1", 00:12:53.755 "superblock": false, 00:12:53.755 "num_base_bdevs": 4, 00:12:53.755 "num_base_bdevs_discovered": 3, 00:12:53.755 "num_base_bdevs_operational": 3, 00:12:53.755 "base_bdevs_list": [ 00:12:53.755 { 00:12:53.755 "name": null, 00:12:53.755 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:53.755 "is_configured": false, 00:12:53.755 "data_offset": 0, 00:12:53.755 "data_size": 65536 00:12:53.755 }, 00:12:53.755 { 00:12:53.755 "name": "BaseBdev2", 00:12:53.755 "uuid": "a6875699-e811-5fd7-b3ef-04ccf6c8d0c1", 00:12:53.755 "is_configured": true, 00:12:53.755 "data_offset": 0, 00:12:53.755 "data_size": 65536 00:12:53.755 }, 00:12:53.755 { 00:12:53.755 "name": "BaseBdev3", 00:12:53.755 "uuid": "4559b304-b699-5e2c-8133-42629be61bbf", 00:12:53.755 "is_configured": true, 00:12:53.755 "data_offset": 0, 00:12:53.755 "data_size": 65536 00:12:53.755 }, 00:12:53.755 { 00:12:53.755 "name": "BaseBdev4", 00:12:53.755 "uuid": "b61c0311-8c43-5e48-82d5-ce72db132f66", 00:12:53.755 "is_configured": true, 00:12:53.755 "data_offset": 0, 00:12:53.755 "data_size": 65536 00:12:53.755 } 00:12:53.755 ] 00:12:53.755 }' 00:12:53.755 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:53.755 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:54.324 160.00 IOPS, 480.00 MiB/s [2024-12-05T19:03:11.883Z] 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:54.324 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:54.324 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:54.324 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:54.324 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:54.324 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:54.324 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.324 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:54.324 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:54.324 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.324 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:54.324 "name": "raid_bdev1", 00:12:54.324 "uuid": "41715d1d-e1e1-4ceb-b212-31d1d4785dab", 00:12:54.324 "strip_size_kb": 0, 00:12:54.324 "state": "online", 00:12:54.324 "raid_level": "raid1", 00:12:54.324 "superblock": false, 00:12:54.324 "num_base_bdevs": 4, 00:12:54.324 "num_base_bdevs_discovered": 3, 00:12:54.324 "num_base_bdevs_operational": 3, 00:12:54.324 "base_bdevs_list": [ 00:12:54.324 { 00:12:54.324 "name": null, 00:12:54.324 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:54.324 "is_configured": false, 00:12:54.324 "data_offset": 0, 00:12:54.324 "data_size": 65536 00:12:54.324 }, 00:12:54.324 { 00:12:54.324 "name": "BaseBdev2", 00:12:54.324 "uuid": "a6875699-e811-5fd7-b3ef-04ccf6c8d0c1", 00:12:54.324 "is_configured": true, 00:12:54.324 "data_offset": 0, 00:12:54.324 "data_size": 65536 00:12:54.324 }, 00:12:54.324 { 00:12:54.324 "name": "BaseBdev3", 00:12:54.324 "uuid": "4559b304-b699-5e2c-8133-42629be61bbf", 00:12:54.324 "is_configured": true, 00:12:54.324 "data_offset": 0, 00:12:54.325 "data_size": 65536 00:12:54.325 }, 00:12:54.325 { 00:12:54.325 "name": "BaseBdev4", 00:12:54.325 "uuid": "b61c0311-8c43-5e48-82d5-ce72db132f66", 00:12:54.325 "is_configured": true, 00:12:54.325 "data_offset": 0, 00:12:54.325 "data_size": 65536 00:12:54.325 } 00:12:54.325 ] 00:12:54.325 }' 00:12:54.325 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:54.325 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:54.325 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:54.325 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:54.325 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:54.325 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:54.325 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:54.325 [2024-12-05 19:03:11.813300] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:54.325 19:03:11 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:54.325 19:03:11 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:54.325 [2024-12-05 19:03:11.848564] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:12:54.325 [2024-12-05 19:03:11.850554] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:54.585 [2024-12-05 19:03:11.958679] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:54.585 [2024-12-05 19:03:11.959811] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:54.845 [2024-12-05 19:03:12.180079] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:54.845 [2024-12-05 19:03:12.180519] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:55.104 [2024-12-05 19:03:12.527230] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:55.364 154.33 IOPS, 463.00 MiB/s [2024-12-05T19:03:12.923Z] [2024-12-05 19:03:12.750515] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:55.364 [2024-12-05 19:03:12.751206] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:55.364 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:55.364 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:55.364 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:55.364 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:55.364 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:55.364 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.364 19:03:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.364 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.364 19:03:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:55.364 19:03:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.364 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:55.364 "name": "raid_bdev1", 00:12:55.364 "uuid": "41715d1d-e1e1-4ceb-b212-31d1d4785dab", 00:12:55.364 "strip_size_kb": 0, 00:12:55.364 "state": "online", 00:12:55.364 "raid_level": "raid1", 00:12:55.364 "superblock": false, 00:12:55.364 "num_base_bdevs": 4, 00:12:55.364 "num_base_bdevs_discovered": 4, 00:12:55.364 "num_base_bdevs_operational": 4, 00:12:55.364 "process": { 00:12:55.364 "type": "rebuild", 00:12:55.364 "target": "spare", 00:12:55.364 "progress": { 00:12:55.364 "blocks": 10240, 00:12:55.364 "percent": 15 00:12:55.364 } 00:12:55.364 }, 00:12:55.364 "base_bdevs_list": [ 00:12:55.364 { 00:12:55.364 "name": "spare", 00:12:55.364 "uuid": "dab5acb0-6f3c-5bc5-86e7-432eb51dcf7d", 00:12:55.364 "is_configured": true, 00:12:55.364 "data_offset": 0, 00:12:55.364 "data_size": 65536 00:12:55.364 }, 00:12:55.364 { 00:12:55.364 "name": "BaseBdev2", 00:12:55.364 "uuid": "a6875699-e811-5fd7-b3ef-04ccf6c8d0c1", 00:12:55.364 "is_configured": true, 00:12:55.364 "data_offset": 0, 00:12:55.364 "data_size": 65536 00:12:55.364 }, 00:12:55.364 { 00:12:55.364 "name": "BaseBdev3", 00:12:55.364 "uuid": "4559b304-b699-5e2c-8133-42629be61bbf", 00:12:55.364 "is_configured": true, 00:12:55.364 "data_offset": 0, 00:12:55.364 "data_size": 65536 00:12:55.364 }, 00:12:55.364 { 00:12:55.364 "name": "BaseBdev4", 00:12:55.364 "uuid": "b61c0311-8c43-5e48-82d5-ce72db132f66", 00:12:55.364 "is_configured": true, 00:12:55.364 "data_offset": 0, 00:12:55.364 "data_size": 65536 00:12:55.364 } 00:12:55.364 ] 00:12:55.364 }' 00:12:55.364 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:55.623 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:55.623 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:55.623 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:55.623 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:55.623 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:55.623 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:55.623 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:55.623 19:03:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:55.623 19:03:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.623 19:03:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:55.623 [2024-12-05 19:03:13.001986] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:55.623 [2024-12-05 19:03:13.107357] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:55.883 [2024-12-05 19:03:13.209351] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002870 00:12:55.883 [2024-12-05 19:03:13.209396] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002a10 00:12:55.883 [2024-12-05 19:03:13.211078] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:55.883 "name": "raid_bdev1", 00:12:55.883 "uuid": "41715d1d-e1e1-4ceb-b212-31d1d4785dab", 00:12:55.883 "strip_size_kb": 0, 00:12:55.883 "state": "online", 00:12:55.883 "raid_level": "raid1", 00:12:55.883 "superblock": false, 00:12:55.883 "num_base_bdevs": 4, 00:12:55.883 "num_base_bdevs_discovered": 3, 00:12:55.883 "num_base_bdevs_operational": 3, 00:12:55.883 "process": { 00:12:55.883 "type": "rebuild", 00:12:55.883 "target": "spare", 00:12:55.883 "progress": { 00:12:55.883 "blocks": 14336, 00:12:55.883 "percent": 21 00:12:55.883 } 00:12:55.883 }, 00:12:55.883 "base_bdevs_list": [ 00:12:55.883 { 00:12:55.883 "name": "spare", 00:12:55.883 "uuid": "dab5acb0-6f3c-5bc5-86e7-432eb51dcf7d", 00:12:55.883 "is_configured": true, 00:12:55.883 "data_offset": 0, 00:12:55.883 "data_size": 65536 00:12:55.883 }, 00:12:55.883 { 00:12:55.883 "name": null, 00:12:55.883 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.883 "is_configured": false, 00:12:55.883 "data_offset": 0, 00:12:55.883 "data_size": 65536 00:12:55.883 }, 00:12:55.883 { 00:12:55.883 "name": "BaseBdev3", 00:12:55.883 "uuid": "4559b304-b699-5e2c-8133-42629be61bbf", 00:12:55.883 "is_configured": true, 00:12:55.883 "data_offset": 0, 00:12:55.883 "data_size": 65536 00:12:55.883 }, 00:12:55.883 { 00:12:55.883 "name": "BaseBdev4", 00:12:55.883 "uuid": "b61c0311-8c43-5e48-82d5-ce72db132f66", 00:12:55.883 "is_configured": true, 00:12:55.883 "data_offset": 0, 00:12:55.883 "data_size": 65536 00:12:55.883 } 00:12:55.883 ] 00:12:55.883 }' 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:55.883 [2024-12-05 19:03:13.319116] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 1 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:55.883 2288 offset_end: 18432 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=383 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:55.883 "name": "raid_bdev1", 00:12:55.883 "uuid": "41715d1d-e1e1-4ceb-b212-31d1d4785dab", 00:12:55.883 "strip_size_kb": 0, 00:12:55.883 "state": "online", 00:12:55.883 "raid_level": "raid1", 00:12:55.883 "superblock": false, 00:12:55.883 "num_base_bdevs": 4, 00:12:55.883 "num_base_bdevs_discovered": 3, 00:12:55.883 "num_base_bdevs_operational": 3, 00:12:55.883 "process": { 00:12:55.883 "type": "rebuild", 00:12:55.883 "target": "spare", 00:12:55.883 "progress": { 00:12:55.883 "blocks": 16384, 00:12:55.883 "percent": 25 00:12:55.883 } 00:12:55.883 }, 00:12:55.883 "base_bdevs_list": [ 00:12:55.883 { 00:12:55.883 "name": "spare", 00:12:55.883 "uuid": "dab5acb0-6f3c-5bc5-86e7-432eb51dcf7d", 00:12:55.883 "is_configured": true, 00:12:55.883 "data_offset": 0, 00:12:55.883 "data_size": 65536 00:12:55.883 }, 00:12:55.883 { 00:12:55.883 "name": null, 00:12:55.883 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:55.883 "is_configured": false, 00:12:55.883 "data_offset": 0, 00:12:55.883 "data_size": 65536 00:12:55.883 }, 00:12:55.883 { 00:12:55.883 "name": "BaseBdev3", 00:12:55.883 "uuid": "4559b304-b699-5e2c-8133-42629be61bbf", 00:12:55.883 "is_configured": true, 00:12:55.883 "data_offset": 0, 00:12:55.883 "data_size": 65536 00:12:55.883 }, 00:12:55.883 { 00:12:55.883 "name": "BaseBdev4", 00:12:55.883 "uuid": "b61c0311-8c43-5e48-82d5-ce72db132f66", 00:12:55.883 "is_configured": true, 00:12:55.883 "data_offset": 0, 00:12:55.883 "data_size": 65536 00:12:55.883 } 00:12:55.883 ] 00:12:55.883 }' 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:55.883 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:56.144 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:56.144 19:03:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:56.144 [2024-12-05 19:03:13.573369] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:12:56.144 130.50 IOPS, 391.50 MiB/s [2024-12-05T19:03:13.703Z] [2024-12-05 19:03:13.699906] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:12:56.403 [2024-12-05 19:03:13.932981] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:12:56.972 19:03:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:56.972 19:03:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:56.972 19:03:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:56.972 19:03:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:56.972 19:03:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:56.972 19:03:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:56.972 19:03:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:56.972 19:03:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:56.972 19:03:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:56.972 19:03:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:56.972 19:03:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:57.232 19:03:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:57.232 "name": "raid_bdev1", 00:12:57.232 "uuid": "41715d1d-e1e1-4ceb-b212-31d1d4785dab", 00:12:57.232 "strip_size_kb": 0, 00:12:57.232 "state": "online", 00:12:57.232 "raid_level": "raid1", 00:12:57.232 "superblock": false, 00:12:57.232 "num_base_bdevs": 4, 00:12:57.232 "num_base_bdevs_discovered": 3, 00:12:57.232 "num_base_bdevs_operational": 3, 00:12:57.232 "process": { 00:12:57.232 "type": "rebuild", 00:12:57.232 "target": "spare", 00:12:57.232 "progress": { 00:12:57.232 "blocks": 32768, 00:12:57.232 "percent": 50 00:12:57.232 } 00:12:57.232 }, 00:12:57.232 "base_bdevs_list": [ 00:12:57.232 { 00:12:57.232 "name": "spare", 00:12:57.232 "uuid": "dab5acb0-6f3c-5bc5-86e7-432eb51dcf7d", 00:12:57.232 "is_configured": true, 00:12:57.232 "data_offset": 0, 00:12:57.232 "data_size": 65536 00:12:57.232 }, 00:12:57.232 { 00:12:57.232 "name": null, 00:12:57.232 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:57.232 "is_configured": false, 00:12:57.232 "data_offset": 0, 00:12:57.232 "data_size": 65536 00:12:57.232 }, 00:12:57.232 { 00:12:57.232 "name": "BaseBdev3", 00:12:57.232 "uuid": "4559b304-b699-5e2c-8133-42629be61bbf", 00:12:57.232 "is_configured": true, 00:12:57.232 "data_offset": 0, 00:12:57.232 "data_size": 65536 00:12:57.232 }, 00:12:57.232 { 00:12:57.232 "name": "BaseBdev4", 00:12:57.232 "uuid": "b61c0311-8c43-5e48-82d5-ce72db132f66", 00:12:57.232 "is_configured": true, 00:12:57.232 "data_offset": 0, 00:12:57.232 "data_size": 65536 00:12:57.232 } 00:12:57.232 ] 00:12:57.232 }' 00:12:57.232 19:03:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:57.233 19:03:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:57.233 19:03:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:57.233 116.40 IOPS, 349.20 MiB/s [2024-12-05T19:03:14.792Z] 19:03:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:57.233 19:03:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:57.492 [2024-12-05 19:03:14.859522] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:12:58.319 103.67 IOPS, 311.00 MiB/s [2024-12-05T19:03:15.878Z] 19:03:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:58.319 "name": "raid_bdev1", 00:12:58.319 "uuid": "41715d1d-e1e1-4ceb-b212-31d1d4785dab", 00:12:58.319 "strip_size_kb": 0, 00:12:58.319 "state": "online", 00:12:58.319 "raid_level": "raid1", 00:12:58.319 "superblock": false, 00:12:58.319 "num_base_bdevs": 4, 00:12:58.319 "num_base_bdevs_discovered": 3, 00:12:58.319 "num_base_bdevs_operational": 3, 00:12:58.319 "process": { 00:12:58.319 "type": "rebuild", 00:12:58.319 "target": "spare", 00:12:58.319 "progress": { 00:12:58.319 "blocks": 53248, 00:12:58.319 "percent": 81 00:12:58.319 } 00:12:58.319 }, 00:12:58.319 "base_bdevs_list": [ 00:12:58.319 { 00:12:58.319 "name": "spare", 00:12:58.319 "uuid": "dab5acb0-6f3c-5bc5-86e7-432eb51dcf7d", 00:12:58.319 "is_configured": true, 00:12:58.319 "data_offset": 0, 00:12:58.319 "data_size": 65536 00:12:58.319 }, 00:12:58.319 { 00:12:58.319 "name": null, 00:12:58.319 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:58.319 "is_configured": false, 00:12:58.319 "data_offset": 0, 00:12:58.319 "data_size": 65536 00:12:58.319 }, 00:12:58.319 { 00:12:58.319 "name": "BaseBdev3", 00:12:58.319 "uuid": "4559b304-b699-5e2c-8133-42629be61bbf", 00:12:58.319 "is_configured": true, 00:12:58.319 "data_offset": 0, 00:12:58.319 "data_size": 65536 00:12:58.319 }, 00:12:58.319 { 00:12:58.319 "name": "BaseBdev4", 00:12:58.319 "uuid": "b61c0311-8c43-5e48-82d5-ce72db132f66", 00:12:58.319 "is_configured": true, 00:12:58.319 "data_offset": 0, 00:12:58.319 "data_size": 65536 00:12:58.319 } 00:12:58.319 ] 00:12:58.319 }' 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:58.319 19:03:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:58.887 [2024-12-05 19:03:16.302830] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:58.887 [2024-12-05 19:03:16.407681] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:58.887 [2024-12-05 19:03:16.409877] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:59.406 94.57 IOPS, 283.71 MiB/s [2024-12-05T19:03:16.965Z] 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:59.406 "name": "raid_bdev1", 00:12:59.406 "uuid": "41715d1d-e1e1-4ceb-b212-31d1d4785dab", 00:12:59.406 "strip_size_kb": 0, 00:12:59.406 "state": "online", 00:12:59.406 "raid_level": "raid1", 00:12:59.406 "superblock": false, 00:12:59.406 "num_base_bdevs": 4, 00:12:59.406 "num_base_bdevs_discovered": 3, 00:12:59.406 "num_base_bdevs_operational": 3, 00:12:59.406 "base_bdevs_list": [ 00:12:59.406 { 00:12:59.406 "name": "spare", 00:12:59.406 "uuid": "dab5acb0-6f3c-5bc5-86e7-432eb51dcf7d", 00:12:59.406 "is_configured": true, 00:12:59.406 "data_offset": 0, 00:12:59.406 "data_size": 65536 00:12:59.406 }, 00:12:59.406 { 00:12:59.406 "name": null, 00:12:59.406 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.406 "is_configured": false, 00:12:59.406 "data_offset": 0, 00:12:59.406 "data_size": 65536 00:12:59.406 }, 00:12:59.406 { 00:12:59.406 "name": "BaseBdev3", 00:12:59.406 "uuid": "4559b304-b699-5e2c-8133-42629be61bbf", 00:12:59.406 "is_configured": true, 00:12:59.406 "data_offset": 0, 00:12:59.406 "data_size": 65536 00:12:59.406 }, 00:12:59.406 { 00:12:59.406 "name": "BaseBdev4", 00:12:59.406 "uuid": "b61c0311-8c43-5e48-82d5-ce72db132f66", 00:12:59.406 "is_configured": true, 00:12:59.406 "data_offset": 0, 00:12:59.406 "data_size": 65536 00:12:59.406 } 00:12:59.406 ] 00:12:59.406 }' 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.406 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:59.406 "name": "raid_bdev1", 00:12:59.406 "uuid": "41715d1d-e1e1-4ceb-b212-31d1d4785dab", 00:12:59.406 "strip_size_kb": 0, 00:12:59.406 "state": "online", 00:12:59.406 "raid_level": "raid1", 00:12:59.406 "superblock": false, 00:12:59.406 "num_base_bdevs": 4, 00:12:59.406 "num_base_bdevs_discovered": 3, 00:12:59.406 "num_base_bdevs_operational": 3, 00:12:59.406 "base_bdevs_list": [ 00:12:59.406 { 00:12:59.406 "name": "spare", 00:12:59.406 "uuid": "dab5acb0-6f3c-5bc5-86e7-432eb51dcf7d", 00:12:59.406 "is_configured": true, 00:12:59.406 "data_offset": 0, 00:12:59.406 "data_size": 65536 00:12:59.406 }, 00:12:59.406 { 00:12:59.406 "name": null, 00:12:59.406 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.406 "is_configured": false, 00:12:59.406 "data_offset": 0, 00:12:59.406 "data_size": 65536 00:12:59.406 }, 00:12:59.406 { 00:12:59.406 "name": "BaseBdev3", 00:12:59.406 "uuid": "4559b304-b699-5e2c-8133-42629be61bbf", 00:12:59.406 "is_configured": true, 00:12:59.406 "data_offset": 0, 00:12:59.406 "data_size": 65536 00:12:59.406 }, 00:12:59.406 { 00:12:59.406 "name": "BaseBdev4", 00:12:59.406 "uuid": "b61c0311-8c43-5e48-82d5-ce72db132f66", 00:12:59.406 "is_configured": true, 00:12:59.406 "data_offset": 0, 00:12:59.406 "data_size": 65536 00:12:59.406 } 00:12:59.406 ] 00:12:59.406 }' 00:12:59.666 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:59.666 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:59.666 19:03:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:59.666 "name": "raid_bdev1", 00:12:59.666 "uuid": "41715d1d-e1e1-4ceb-b212-31d1d4785dab", 00:12:59.666 "strip_size_kb": 0, 00:12:59.666 "state": "online", 00:12:59.666 "raid_level": "raid1", 00:12:59.666 "superblock": false, 00:12:59.666 "num_base_bdevs": 4, 00:12:59.666 "num_base_bdevs_discovered": 3, 00:12:59.666 "num_base_bdevs_operational": 3, 00:12:59.666 "base_bdevs_list": [ 00:12:59.666 { 00:12:59.666 "name": "spare", 00:12:59.666 "uuid": "dab5acb0-6f3c-5bc5-86e7-432eb51dcf7d", 00:12:59.666 "is_configured": true, 00:12:59.666 "data_offset": 0, 00:12:59.666 "data_size": 65536 00:12:59.666 }, 00:12:59.666 { 00:12:59.666 "name": null, 00:12:59.666 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.666 "is_configured": false, 00:12:59.666 "data_offset": 0, 00:12:59.666 "data_size": 65536 00:12:59.666 }, 00:12:59.666 { 00:12:59.666 "name": "BaseBdev3", 00:12:59.666 "uuid": "4559b304-b699-5e2c-8133-42629be61bbf", 00:12:59.666 "is_configured": true, 00:12:59.666 "data_offset": 0, 00:12:59.666 "data_size": 65536 00:12:59.666 }, 00:12:59.666 { 00:12:59.666 "name": "BaseBdev4", 00:12:59.666 "uuid": "b61c0311-8c43-5e48-82d5-ce72db132f66", 00:12:59.666 "is_configured": true, 00:12:59.666 "data_offset": 0, 00:12:59.666 "data_size": 65536 00:12:59.666 } 00:12:59.666 ] 00:12:59.666 }' 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:59.666 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.235 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:00.235 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.235 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.235 [2024-12-05 19:03:17.502404] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:00.235 [2024-12-05 19:03:17.502508] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:00.235 88.12 IOPS, 264.38 MiB/s 00:13:00.235 Latency(us) 00:13:00.235 [2024-12-05T19:03:17.794Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:00.235 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:13:00.235 raid_bdev1 : 8.01 88.18 264.55 0.00 0.00 15888.37 291.55 116304.94 00:13:00.235 [2024-12-05T19:03:17.794Z] =================================================================================================================== 00:13:00.235 [2024-12-05T19:03:17.794Z] Total : 88.18 264.55 0.00 0.00 15888.37 291.55 116304.94 00:13:00.235 [2024-12-05 19:03:17.598609] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:00.235 [2024-12-05 19:03:17.598733] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:00.235 [2024-12-05 19:03:17.598868] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:00.235 [2024-12-05 19:03:17.598923] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:00.235 { 00:13:00.235 "results": [ 00:13:00.235 { 00:13:00.235 "job": "raid_bdev1", 00:13:00.235 "core_mask": "0x1", 00:13:00.235 "workload": "randrw", 00:13:00.235 "percentage": 50, 00:13:00.235 "status": "finished", 00:13:00.235 "queue_depth": 2, 00:13:00.235 "io_size": 3145728, 00:13:00.235 "runtime": 8.006174, 00:13:00.235 "iops": 88.18194558349593, 00:13:00.235 "mibps": 264.5458367504878, 00:13:00.235 "io_failed": 0, 00:13:00.235 "io_timeout": 0, 00:13:00.235 "avg_latency_us": 15888.365100139787, 00:13:00.235 "min_latency_us": 291.54934497816595, 00:13:00.235 "max_latency_us": 116304.93624454149 00:13:00.235 } 00:13:00.235 ], 00:13:00.235 "core_count": 1 00:13:00.235 } 00:13:00.235 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.235 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:00.235 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:13:00.235 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:00.235 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.235 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:00.235 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:00.235 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:00.235 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:13:00.236 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:13:00.236 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:00.236 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:13:00.236 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:00.236 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:00.236 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:00.236 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:00.236 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:00.236 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:00.236 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:13:00.496 /dev/nbd0 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:00.496 1+0 records in 00:13:00.496 1+0 records out 00:13:00.496 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000478191 s, 8.6 MB/s 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@728 -- # continue 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:00.496 19:03:17 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:13:00.755 /dev/nbd1 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:00.755 1+0 records in 00:13:00.755 1+0 records out 00:13:00.755 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00056899 s, 7.2 MB/s 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:00.755 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:01.015 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:13:01.276 /dev/nbd1 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # local i 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@877 -- # break 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:01.276 1+0 records in 00:13:01.276 1+0 records out 00:13:01.276 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000391695 s, 10.5 MB/s 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@890 -- # size=4096 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@893 -- # return 0 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:01.276 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:01.535 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:01.535 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:01.535 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:01.535 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:01.535 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:01.535 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:01.535 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:01.535 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:01.535 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:01.535 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:01.535 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:01.535 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:01.535 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:01.535 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:01.535 19:03:18 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 87494 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # '[' -z 87494 ']' 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@958 -- # kill -0 87494 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # uname 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87494 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87494' 00:13:01.794 killing process with pid 87494 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@973 -- # kill 87494 00:13:01.794 Received shutdown signal, test time was about 9.645109 seconds 00:13:01.794 00:13:01.794 Latency(us) 00:13:01.794 [2024-12-05T19:03:19.353Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:01.794 [2024-12-05T19:03:19.353Z] =================================================================================================================== 00:13:01.794 [2024-12-05T19:03:19.353Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:01.794 [2024-12-05 19:03:19.231029] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:01.794 19:03:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@978 -- # wait 87494 00:13:01.794 [2024-12-05 19:03:19.275248] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:13:02.053 00:13:02.053 real 0m11.614s 00:13:02.053 user 0m15.103s 00:13:02.053 sys 0m1.785s 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:02.053 ************************************ 00:13:02.053 END TEST raid_rebuild_test_io 00:13:02.053 ************************************ 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.053 19:03:19 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 4 true true true 00:13:02.053 19:03:19 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:13:02.053 19:03:19 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:02.053 19:03:19 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:02.053 ************************************ 00:13:02.053 START TEST raid_rebuild_test_sb_io 00:13:02.053 ************************************ 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 4 true true true 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=87820 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 87820 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # '[' -z 87820 ']' 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:02.053 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:02.053 19:03:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.313 [2024-12-05 19:03:19.655256] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:13:02.313 [2024-12-05 19:03:19.655458] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:13:02.313 Zero copy mechanism will not be used. 00:13:02.313 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87820 ] 00:13:02.313 [2024-12-05 19:03:19.811473] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:02.313 [2024-12-05 19:03:19.836644] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:02.573 [2024-12-05 19:03:19.880162] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:02.573 [2024-12-05 19:03:19.880276] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # return 0 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.144 BaseBdev1_malloc 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.144 [2024-12-05 19:03:20.488666] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:03.144 [2024-12-05 19:03:20.488732] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:03.144 [2024-12-05 19:03:20.488764] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:03.144 [2024-12-05 19:03:20.488783] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:03.144 [2024-12-05 19:03:20.490832] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:03.144 [2024-12-05 19:03:20.490866] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:03.144 BaseBdev1 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.144 BaseBdev2_malloc 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.144 [2024-12-05 19:03:20.517332] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:03.144 [2024-12-05 19:03:20.517381] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:03.144 [2024-12-05 19:03:20.517403] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:03.144 [2024-12-05 19:03:20.517412] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:03.144 [2024-12-05 19:03:20.519536] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:03.144 [2024-12-05 19:03:20.519577] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:03.144 BaseBdev2 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.144 BaseBdev3_malloc 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.144 [2024-12-05 19:03:20.545887] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:03.144 [2024-12-05 19:03:20.545963] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:03.144 [2024-12-05 19:03:20.545985] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:03.144 [2024-12-05 19:03:20.545994] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:03.144 [2024-12-05 19:03:20.548098] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:03.144 [2024-12-05 19:03:20.548134] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:03.144 BaseBdev3 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.144 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.144 BaseBdev4_malloc 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.145 [2024-12-05 19:03:20.589862] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:13:03.145 [2024-12-05 19:03:20.589913] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:03.145 [2024-12-05 19:03:20.589937] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:03.145 [2024-12-05 19:03:20.589947] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:03.145 [2024-12-05 19:03:20.592169] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:03.145 [2024-12-05 19:03:20.592208] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:13:03.145 BaseBdev4 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.145 spare_malloc 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.145 spare_delay 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.145 [2024-12-05 19:03:20.630453] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:03.145 [2024-12-05 19:03:20.630539] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:03.145 [2024-12-05 19:03:20.630560] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:13:03.145 [2024-12-05 19:03:20.630569] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:03.145 [2024-12-05 19:03:20.632677] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:03.145 [2024-12-05 19:03:20.632715] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:03.145 spare 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.145 [2024-12-05 19:03:20.642502] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:03.145 [2024-12-05 19:03:20.644328] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:03.145 [2024-12-05 19:03:20.644391] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:03.145 [2024-12-05 19:03:20.644439] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:03.145 [2024-12-05 19:03:20.644617] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:03.145 [2024-12-05 19:03:20.644628] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:03.145 [2024-12-05 19:03:20.644866] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:13:03.145 [2024-12-05 19:03:20.645003] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:03.145 [2024-12-05 19:03:20.645015] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:03.145 [2024-12-05 19:03:20.645127] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:03.145 "name": "raid_bdev1", 00:13:03.145 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:03.145 "strip_size_kb": 0, 00:13:03.145 "state": "online", 00:13:03.145 "raid_level": "raid1", 00:13:03.145 "superblock": true, 00:13:03.145 "num_base_bdevs": 4, 00:13:03.145 "num_base_bdevs_discovered": 4, 00:13:03.145 "num_base_bdevs_operational": 4, 00:13:03.145 "base_bdevs_list": [ 00:13:03.145 { 00:13:03.145 "name": "BaseBdev1", 00:13:03.145 "uuid": "85968df5-90e9-518f-9df1-2424803e68ff", 00:13:03.145 "is_configured": true, 00:13:03.145 "data_offset": 2048, 00:13:03.145 "data_size": 63488 00:13:03.145 }, 00:13:03.145 { 00:13:03.145 "name": "BaseBdev2", 00:13:03.145 "uuid": "98bfe603-f4e4-5d67-861e-2dbae99cdc34", 00:13:03.145 "is_configured": true, 00:13:03.145 "data_offset": 2048, 00:13:03.145 "data_size": 63488 00:13:03.145 }, 00:13:03.145 { 00:13:03.145 "name": "BaseBdev3", 00:13:03.145 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:03.145 "is_configured": true, 00:13:03.145 "data_offset": 2048, 00:13:03.145 "data_size": 63488 00:13:03.145 }, 00:13:03.145 { 00:13:03.145 "name": "BaseBdev4", 00:13:03.145 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:03.145 "is_configured": true, 00:13:03.145 "data_offset": 2048, 00:13:03.145 "data_size": 63488 00:13:03.145 } 00:13:03.145 ] 00:13:03.145 }' 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:03.145 19:03:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.715 [2024-12-05 19:03:21.105974] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.715 [2024-12-05 19:03:21.197489] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:03.715 "name": "raid_bdev1", 00:13:03.715 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:03.715 "strip_size_kb": 0, 00:13:03.715 "state": "online", 00:13:03.715 "raid_level": "raid1", 00:13:03.715 "superblock": true, 00:13:03.715 "num_base_bdevs": 4, 00:13:03.715 "num_base_bdevs_discovered": 3, 00:13:03.715 "num_base_bdevs_operational": 3, 00:13:03.715 "base_bdevs_list": [ 00:13:03.715 { 00:13:03.715 "name": null, 00:13:03.715 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:03.715 "is_configured": false, 00:13:03.715 "data_offset": 0, 00:13:03.715 "data_size": 63488 00:13:03.715 }, 00:13:03.715 { 00:13:03.715 "name": "BaseBdev2", 00:13:03.715 "uuid": "98bfe603-f4e4-5d67-861e-2dbae99cdc34", 00:13:03.715 "is_configured": true, 00:13:03.715 "data_offset": 2048, 00:13:03.715 "data_size": 63488 00:13:03.715 }, 00:13:03.715 { 00:13:03.715 "name": "BaseBdev3", 00:13:03.715 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:03.715 "is_configured": true, 00:13:03.715 "data_offset": 2048, 00:13:03.715 "data_size": 63488 00:13:03.715 }, 00:13:03.715 { 00:13:03.715 "name": "BaseBdev4", 00:13:03.715 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:03.715 "is_configured": true, 00:13:03.715 "data_offset": 2048, 00:13:03.715 "data_size": 63488 00:13:03.715 } 00:13:03.715 ] 00:13:03.715 }' 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:03.715 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.988 [2024-12-05 19:03:21.283395] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:13:03.988 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:03.988 Zero copy mechanism will not be used. 00:13:03.988 Running I/O for 60 seconds... 00:13:04.270 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:04.270 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:04.270 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.270 [2024-12-05 19:03:21.639063] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:04.270 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:04.270 19:03:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:04.270 [2024-12-05 19:03:21.686947] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002940 00:13:04.270 [2024-12-05 19:03:21.688942] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:04.270 [2024-12-05 19:03:21.807871] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:04.270 [2024-12-05 19:03:21.809036] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:04.542 [2024-12-05 19:03:22.010936] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:04.542 [2024-12-05 19:03:22.011246] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:04.801 [2024-12-05 19:03:22.260653] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:05.061 140.00 IOPS, 420.00 MiB/s [2024-12-05T19:03:22.620Z] [2024-12-05 19:03:22.389585] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:05.061 [2024-12-05 19:03:22.389819] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:05.061 [2024-12-05 19:03:22.606937] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:05.322 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:05.322 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:05.322 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:05.322 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:05.322 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:05.322 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.322 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:05.322 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.322 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:05.322 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.322 [2024-12-05 19:03:22.709441] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:05.322 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:05.322 "name": "raid_bdev1", 00:13:05.322 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:05.322 "strip_size_kb": 0, 00:13:05.322 "state": "online", 00:13:05.322 "raid_level": "raid1", 00:13:05.322 "superblock": true, 00:13:05.322 "num_base_bdevs": 4, 00:13:05.322 "num_base_bdevs_discovered": 4, 00:13:05.322 "num_base_bdevs_operational": 4, 00:13:05.322 "process": { 00:13:05.322 "type": "rebuild", 00:13:05.322 "target": "spare", 00:13:05.322 "progress": { 00:13:05.322 "blocks": 14336, 00:13:05.322 "percent": 22 00:13:05.322 } 00:13:05.322 }, 00:13:05.322 "base_bdevs_list": [ 00:13:05.322 { 00:13:05.322 "name": "spare", 00:13:05.322 "uuid": "41cc6ffc-af87-5de8-a83b-08845eb57c3f", 00:13:05.322 "is_configured": true, 00:13:05.322 "data_offset": 2048, 00:13:05.322 "data_size": 63488 00:13:05.322 }, 00:13:05.322 { 00:13:05.322 "name": "BaseBdev2", 00:13:05.322 "uuid": "98bfe603-f4e4-5d67-861e-2dbae99cdc34", 00:13:05.322 "is_configured": true, 00:13:05.322 "data_offset": 2048, 00:13:05.322 "data_size": 63488 00:13:05.322 }, 00:13:05.323 { 00:13:05.323 "name": "BaseBdev3", 00:13:05.323 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:05.323 "is_configured": true, 00:13:05.323 "data_offset": 2048, 00:13:05.323 "data_size": 63488 00:13:05.323 }, 00:13:05.323 { 00:13:05.323 "name": "BaseBdev4", 00:13:05.323 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:05.323 "is_configured": true, 00:13:05.323 "data_offset": 2048, 00:13:05.323 "data_size": 63488 00:13:05.323 } 00:13:05.323 ] 00:13:05.323 }' 00:13:05.323 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:05.323 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:05.323 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:05.323 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:05.323 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:05.323 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.323 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:05.323 [2024-12-05 19:03:22.808049] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:05.323 [2024-12-05 19:03:22.817659] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:05.583 [2024-12-05 19:03:22.924629] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:05.583 [2024-12-05 19:03:22.928906] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:05.583 [2024-12-05 19:03:22.929016] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:05.583 [2024-12-05 19:03:22.929032] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:05.583 [2024-12-05 19:03:22.948072] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002870 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:05.583 19:03:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:05.583 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:05.583 "name": "raid_bdev1", 00:13:05.583 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:05.583 "strip_size_kb": 0, 00:13:05.583 "state": "online", 00:13:05.583 "raid_level": "raid1", 00:13:05.583 "superblock": true, 00:13:05.583 "num_base_bdevs": 4, 00:13:05.583 "num_base_bdevs_discovered": 3, 00:13:05.583 "num_base_bdevs_operational": 3, 00:13:05.583 "base_bdevs_list": [ 00:13:05.583 { 00:13:05.583 "name": null, 00:13:05.583 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:05.583 "is_configured": false, 00:13:05.583 "data_offset": 0, 00:13:05.583 "data_size": 63488 00:13:05.583 }, 00:13:05.583 { 00:13:05.583 "name": "BaseBdev2", 00:13:05.583 "uuid": "98bfe603-f4e4-5d67-861e-2dbae99cdc34", 00:13:05.583 "is_configured": true, 00:13:05.583 "data_offset": 2048, 00:13:05.583 "data_size": 63488 00:13:05.583 }, 00:13:05.583 { 00:13:05.583 "name": "BaseBdev3", 00:13:05.583 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:05.583 "is_configured": true, 00:13:05.583 "data_offset": 2048, 00:13:05.583 "data_size": 63488 00:13:05.583 }, 00:13:05.583 { 00:13:05.583 "name": "BaseBdev4", 00:13:05.583 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:05.583 "is_configured": true, 00:13:05.583 "data_offset": 2048, 00:13:05.583 "data_size": 63488 00:13:05.583 } 00:13:05.583 ] 00:13:05.583 }' 00:13:05.583 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:05.583 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.101 142.50 IOPS, 427.50 MiB/s [2024-12-05T19:03:23.660Z] 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:06.101 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:06.101 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:06.101 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:06.101 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:06.101 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:06.101 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:06.102 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.102 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.102 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.102 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:06.102 "name": "raid_bdev1", 00:13:06.102 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:06.102 "strip_size_kb": 0, 00:13:06.102 "state": "online", 00:13:06.102 "raid_level": "raid1", 00:13:06.102 "superblock": true, 00:13:06.102 "num_base_bdevs": 4, 00:13:06.102 "num_base_bdevs_discovered": 3, 00:13:06.102 "num_base_bdevs_operational": 3, 00:13:06.102 "base_bdevs_list": [ 00:13:06.102 { 00:13:06.102 "name": null, 00:13:06.102 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:06.102 "is_configured": false, 00:13:06.102 "data_offset": 0, 00:13:06.102 "data_size": 63488 00:13:06.102 }, 00:13:06.102 { 00:13:06.102 "name": "BaseBdev2", 00:13:06.102 "uuid": "98bfe603-f4e4-5d67-861e-2dbae99cdc34", 00:13:06.102 "is_configured": true, 00:13:06.102 "data_offset": 2048, 00:13:06.102 "data_size": 63488 00:13:06.102 }, 00:13:06.102 { 00:13:06.102 "name": "BaseBdev3", 00:13:06.102 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:06.102 "is_configured": true, 00:13:06.102 "data_offset": 2048, 00:13:06.102 "data_size": 63488 00:13:06.102 }, 00:13:06.102 { 00:13:06.102 "name": "BaseBdev4", 00:13:06.102 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:06.102 "is_configured": true, 00:13:06.102 "data_offset": 2048, 00:13:06.102 "data_size": 63488 00:13:06.102 } 00:13:06.102 ] 00:13:06.102 }' 00:13:06.102 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:06.102 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:06.102 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:06.102 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:06.102 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:06.102 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:06.102 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.102 [2024-12-05 19:03:23.568810] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:06.102 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:06.102 19:03:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:06.102 [2024-12-05 19:03:23.613564] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:13:06.102 [2024-12-05 19:03:23.615559] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:06.361 [2024-12-05 19:03:23.724156] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:06.361 [2024-12-05 19:03:23.725310] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:06.620 [2024-12-05 19:03:23.931860] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:06.620 [2024-12-05 19:03:23.932460] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:06.879 [2024-12-05 19:03:24.279871] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:07.139 145.33 IOPS, 436.00 MiB/s [2024-12-05T19:03:24.698Z] [2024-12-05 19:03:24.500767] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:07.139 [2024-12-05 19:03:24.501557] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:07.139 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:07.139 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:07.139 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:07.139 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:07.139 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:07.139 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.139 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:07.139 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:07.139 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:07.140 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:07.140 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:07.140 "name": "raid_bdev1", 00:13:07.140 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:07.140 "strip_size_kb": 0, 00:13:07.140 "state": "online", 00:13:07.140 "raid_level": "raid1", 00:13:07.140 "superblock": true, 00:13:07.140 "num_base_bdevs": 4, 00:13:07.140 "num_base_bdevs_discovered": 4, 00:13:07.140 "num_base_bdevs_operational": 4, 00:13:07.140 "process": { 00:13:07.140 "type": "rebuild", 00:13:07.140 "target": "spare", 00:13:07.140 "progress": { 00:13:07.140 "blocks": 10240, 00:13:07.140 "percent": 16 00:13:07.140 } 00:13:07.140 }, 00:13:07.140 "base_bdevs_list": [ 00:13:07.140 { 00:13:07.140 "name": "spare", 00:13:07.140 "uuid": "41cc6ffc-af87-5de8-a83b-08845eb57c3f", 00:13:07.140 "is_configured": true, 00:13:07.140 "data_offset": 2048, 00:13:07.140 "data_size": 63488 00:13:07.140 }, 00:13:07.140 { 00:13:07.140 "name": "BaseBdev2", 00:13:07.140 "uuid": "98bfe603-f4e4-5d67-861e-2dbae99cdc34", 00:13:07.140 "is_configured": true, 00:13:07.140 "data_offset": 2048, 00:13:07.140 "data_size": 63488 00:13:07.140 }, 00:13:07.140 { 00:13:07.140 "name": "BaseBdev3", 00:13:07.140 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:07.140 "is_configured": true, 00:13:07.140 "data_offset": 2048, 00:13:07.140 "data_size": 63488 00:13:07.140 }, 00:13:07.140 { 00:13:07.140 "name": "BaseBdev4", 00:13:07.140 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:07.140 "is_configured": true, 00:13:07.140 "data_offset": 2048, 00:13:07.140 "data_size": 63488 00:13:07.140 } 00:13:07.140 ] 00:13:07.140 }' 00:13:07.140 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:07.400 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:07.400 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:07.400 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:07.400 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:13:07.400 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:13:07.400 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:13:07.400 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:13:07.400 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:07.400 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:13:07.400 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:07.400 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:07.400 19:03:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:07.400 [2024-12-05 19:03:24.759600] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:07.400 [2024-12-05 19:03:24.861286] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:07.661 [2024-12-05 19:03:25.079748] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002870 00:13:07.661 [2024-12-05 19:03:25.079787] bdev_raid.c:1974:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002a10 00:13:07.661 [2024-12-05 19:03:25.088601] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:07.661 "name": "raid_bdev1", 00:13:07.661 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:07.661 "strip_size_kb": 0, 00:13:07.661 "state": "online", 00:13:07.661 "raid_level": "raid1", 00:13:07.661 "superblock": true, 00:13:07.661 "num_base_bdevs": 4, 00:13:07.661 "num_base_bdevs_discovered": 3, 00:13:07.661 "num_base_bdevs_operational": 3, 00:13:07.661 "process": { 00:13:07.661 "type": "rebuild", 00:13:07.661 "target": "spare", 00:13:07.661 "progress": { 00:13:07.661 "blocks": 14336, 00:13:07.661 "percent": 22 00:13:07.661 } 00:13:07.661 }, 00:13:07.661 "base_bdevs_list": [ 00:13:07.661 { 00:13:07.661 "name": "spare", 00:13:07.661 "uuid": "41cc6ffc-af87-5de8-a83b-08845eb57c3f", 00:13:07.661 "is_configured": true, 00:13:07.661 "data_offset": 2048, 00:13:07.661 "data_size": 63488 00:13:07.661 }, 00:13:07.661 { 00:13:07.661 "name": null, 00:13:07.661 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.661 "is_configured": false, 00:13:07.661 "data_offset": 0, 00:13:07.661 "data_size": 63488 00:13:07.661 }, 00:13:07.661 { 00:13:07.661 "name": "BaseBdev3", 00:13:07.661 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:07.661 "is_configured": true, 00:13:07.661 "data_offset": 2048, 00:13:07.661 "data_size": 63488 00:13:07.661 }, 00:13:07.661 { 00:13:07.661 "name": "BaseBdev4", 00:13:07.661 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:07.661 "is_configured": true, 00:13:07.661 "data_offset": 2048, 00:13:07.661 "data_size": 63488 00:13:07.661 } 00:13:07.661 ] 00:13:07.661 }' 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:07.661 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:07.661 [2024-12-05 19:03:25.203740] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:07.661 [2024-12-05 19:03:25.203977] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:07.920 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:07.920 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=395 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:07.921 "name": "raid_bdev1", 00:13:07.921 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:07.921 "strip_size_kb": 0, 00:13:07.921 "state": "online", 00:13:07.921 "raid_level": "raid1", 00:13:07.921 "superblock": true, 00:13:07.921 "num_base_bdevs": 4, 00:13:07.921 "num_base_bdevs_discovered": 3, 00:13:07.921 "num_base_bdevs_operational": 3, 00:13:07.921 "process": { 00:13:07.921 "type": "rebuild", 00:13:07.921 "target": "spare", 00:13:07.921 "progress": { 00:13:07.921 "blocks": 16384, 00:13:07.921 "percent": 25 00:13:07.921 } 00:13:07.921 }, 00:13:07.921 "base_bdevs_list": [ 00:13:07.921 { 00:13:07.921 "name": "spare", 00:13:07.921 "uuid": "41cc6ffc-af87-5de8-a83b-08845eb57c3f", 00:13:07.921 "is_configured": true, 00:13:07.921 "data_offset": 2048, 00:13:07.921 "data_size": 63488 00:13:07.921 }, 00:13:07.921 { 00:13:07.921 "name": null, 00:13:07.921 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.921 "is_configured": false, 00:13:07.921 "data_offset": 0, 00:13:07.921 "data_size": 63488 00:13:07.921 }, 00:13:07.921 { 00:13:07.921 "name": "BaseBdev3", 00:13:07.921 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:07.921 "is_configured": true, 00:13:07.921 "data_offset": 2048, 00:13:07.921 "data_size": 63488 00:13:07.921 }, 00:13:07.921 { 00:13:07.921 "name": "BaseBdev4", 00:13:07.921 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:07.921 "is_configured": true, 00:13:07.921 "data_offset": 2048, 00:13:07.921 "data_size": 63488 00:13:07.921 } 00:13:07.921 ] 00:13:07.921 }' 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:07.921 123.75 IOPS, 371.25 MiB/s [2024-12-05T19:03:25.480Z] 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:07.921 19:03:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:07.921 [2024-12-05 19:03:25.431736] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:13:07.921 [2024-12-05 19:03:25.432625] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:13:08.489 [2024-12-05 19:03:25.985211] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:13:08.748 118.60 IOPS, 355.80 MiB/s [2024-12-05T19:03:26.307Z] [2024-12-05 19:03:26.296490] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:13:08.748 [2024-12-05 19:03:26.297255] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:09.008 "name": "raid_bdev1", 00:13:09.008 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:09.008 "strip_size_kb": 0, 00:13:09.008 "state": "online", 00:13:09.008 "raid_level": "raid1", 00:13:09.008 "superblock": true, 00:13:09.008 "num_base_bdevs": 4, 00:13:09.008 "num_base_bdevs_discovered": 3, 00:13:09.008 "num_base_bdevs_operational": 3, 00:13:09.008 "process": { 00:13:09.008 "type": "rebuild", 00:13:09.008 "target": "spare", 00:13:09.008 "progress": { 00:13:09.008 "blocks": 32768, 00:13:09.008 "percent": 51 00:13:09.008 } 00:13:09.008 }, 00:13:09.008 "base_bdevs_list": [ 00:13:09.008 { 00:13:09.008 "name": "spare", 00:13:09.008 "uuid": "41cc6ffc-af87-5de8-a83b-08845eb57c3f", 00:13:09.008 "is_configured": true, 00:13:09.008 "data_offset": 2048, 00:13:09.008 "data_size": 63488 00:13:09.008 }, 00:13:09.008 { 00:13:09.008 "name": null, 00:13:09.008 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:09.008 "is_configured": false, 00:13:09.008 "data_offset": 0, 00:13:09.008 "data_size": 63488 00:13:09.008 }, 00:13:09.008 { 00:13:09.008 "name": "BaseBdev3", 00:13:09.008 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:09.008 "is_configured": true, 00:13:09.008 "data_offset": 2048, 00:13:09.008 "data_size": 63488 00:13:09.008 }, 00:13:09.008 { 00:13:09.008 "name": "BaseBdev4", 00:13:09.008 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:09.008 "is_configured": true, 00:13:09.008 "data_offset": 2048, 00:13:09.008 "data_size": 63488 00:13:09.008 } 00:13:09.008 ] 00:13:09.008 }' 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:09.008 19:03:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:10.210 106.67 IOPS, 320.00 MiB/s [2024-12-05T19:03:27.769Z] 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:10.210 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:10.210 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:10.210 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:10.210 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:10.210 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:10.210 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:10.210 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.210 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:10.210 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:10.210 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:10.210 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:10.210 "name": "raid_bdev1", 00:13:10.210 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:10.210 "strip_size_kb": 0, 00:13:10.210 "state": "online", 00:13:10.210 "raid_level": "raid1", 00:13:10.210 "superblock": true, 00:13:10.210 "num_base_bdevs": 4, 00:13:10.210 "num_base_bdevs_discovered": 3, 00:13:10.210 "num_base_bdevs_operational": 3, 00:13:10.210 "process": { 00:13:10.210 "type": "rebuild", 00:13:10.210 "target": "spare", 00:13:10.210 "progress": { 00:13:10.210 "blocks": 53248, 00:13:10.210 "percent": 83 00:13:10.210 } 00:13:10.210 }, 00:13:10.210 "base_bdevs_list": [ 00:13:10.210 { 00:13:10.210 "name": "spare", 00:13:10.210 "uuid": "41cc6ffc-af87-5de8-a83b-08845eb57c3f", 00:13:10.210 "is_configured": true, 00:13:10.210 "data_offset": 2048, 00:13:10.210 "data_size": 63488 00:13:10.210 }, 00:13:10.210 { 00:13:10.210 "name": null, 00:13:10.210 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:10.210 "is_configured": false, 00:13:10.210 "data_offset": 0, 00:13:10.210 "data_size": 63488 00:13:10.210 }, 00:13:10.210 { 00:13:10.210 "name": "BaseBdev3", 00:13:10.210 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:10.210 "is_configured": true, 00:13:10.210 "data_offset": 2048, 00:13:10.210 "data_size": 63488 00:13:10.210 }, 00:13:10.210 { 00:13:10.210 "name": "BaseBdev4", 00:13:10.210 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:10.210 "is_configured": true, 00:13:10.210 "data_offset": 2048, 00:13:10.210 "data_size": 63488 00:13:10.210 } 00:13:10.210 ] 00:13:10.210 }' 00:13:10.210 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:10.210 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:10.211 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:10.211 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:10.211 19:03:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:10.781 [2024-12-05 19:03:28.046588] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:10.781 [2024-12-05 19:03:28.146447] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:10.781 [2024-12-05 19:03:28.154072] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:11.350 97.00 IOPS, 291.00 MiB/s [2024-12-05T19:03:28.910Z] 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:11.351 "name": "raid_bdev1", 00:13:11.351 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:11.351 "strip_size_kb": 0, 00:13:11.351 "state": "online", 00:13:11.351 "raid_level": "raid1", 00:13:11.351 "superblock": true, 00:13:11.351 "num_base_bdevs": 4, 00:13:11.351 "num_base_bdevs_discovered": 3, 00:13:11.351 "num_base_bdevs_operational": 3, 00:13:11.351 "base_bdevs_list": [ 00:13:11.351 { 00:13:11.351 "name": "spare", 00:13:11.351 "uuid": "41cc6ffc-af87-5de8-a83b-08845eb57c3f", 00:13:11.351 "is_configured": true, 00:13:11.351 "data_offset": 2048, 00:13:11.351 "data_size": 63488 00:13:11.351 }, 00:13:11.351 { 00:13:11.351 "name": null, 00:13:11.351 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.351 "is_configured": false, 00:13:11.351 "data_offset": 0, 00:13:11.351 "data_size": 63488 00:13:11.351 }, 00:13:11.351 { 00:13:11.351 "name": "BaseBdev3", 00:13:11.351 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:11.351 "is_configured": true, 00:13:11.351 "data_offset": 2048, 00:13:11.351 "data_size": 63488 00:13:11.351 }, 00:13:11.351 { 00:13:11.351 "name": "BaseBdev4", 00:13:11.351 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:11.351 "is_configured": true, 00:13:11.351 "data_offset": 2048, 00:13:11.351 "data_size": 63488 00:13:11.351 } 00:13:11.351 ] 00:13:11.351 }' 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:11.351 "name": "raid_bdev1", 00:13:11.351 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:11.351 "strip_size_kb": 0, 00:13:11.351 "state": "online", 00:13:11.351 "raid_level": "raid1", 00:13:11.351 "superblock": true, 00:13:11.351 "num_base_bdevs": 4, 00:13:11.351 "num_base_bdevs_discovered": 3, 00:13:11.351 "num_base_bdevs_operational": 3, 00:13:11.351 "base_bdevs_list": [ 00:13:11.351 { 00:13:11.351 "name": "spare", 00:13:11.351 "uuid": "41cc6ffc-af87-5de8-a83b-08845eb57c3f", 00:13:11.351 "is_configured": true, 00:13:11.351 "data_offset": 2048, 00:13:11.351 "data_size": 63488 00:13:11.351 }, 00:13:11.351 { 00:13:11.351 "name": null, 00:13:11.351 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.351 "is_configured": false, 00:13:11.351 "data_offset": 0, 00:13:11.351 "data_size": 63488 00:13:11.351 }, 00:13:11.351 { 00:13:11.351 "name": "BaseBdev3", 00:13:11.351 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:11.351 "is_configured": true, 00:13:11.351 "data_offset": 2048, 00:13:11.351 "data_size": 63488 00:13:11.351 }, 00:13:11.351 { 00:13:11.351 "name": "BaseBdev4", 00:13:11.351 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:11.351 "is_configured": true, 00:13:11.351 "data_offset": 2048, 00:13:11.351 "data_size": 63488 00:13:11.351 } 00:13:11.351 ] 00:13:11.351 }' 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:11.351 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:11.611 19:03:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:11.611 "name": "raid_bdev1", 00:13:11.611 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:11.611 "strip_size_kb": 0, 00:13:11.611 "state": "online", 00:13:11.611 "raid_level": "raid1", 00:13:11.611 "superblock": true, 00:13:11.611 "num_base_bdevs": 4, 00:13:11.611 "num_base_bdevs_discovered": 3, 00:13:11.611 "num_base_bdevs_operational": 3, 00:13:11.611 "base_bdevs_list": [ 00:13:11.611 { 00:13:11.611 "name": "spare", 00:13:11.611 "uuid": "41cc6ffc-af87-5de8-a83b-08845eb57c3f", 00:13:11.611 "is_configured": true, 00:13:11.611 "data_offset": 2048, 00:13:11.611 "data_size": 63488 00:13:11.611 }, 00:13:11.611 { 00:13:11.611 "name": null, 00:13:11.611 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.611 "is_configured": false, 00:13:11.611 "data_offset": 0, 00:13:11.611 "data_size": 63488 00:13:11.611 }, 00:13:11.611 { 00:13:11.611 "name": "BaseBdev3", 00:13:11.611 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:11.611 "is_configured": true, 00:13:11.611 "data_offset": 2048, 00:13:11.611 "data_size": 63488 00:13:11.611 }, 00:13:11.611 { 00:13:11.611 "name": "BaseBdev4", 00:13:11.611 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:11.611 "is_configured": true, 00:13:11.611 "data_offset": 2048, 00:13:11.611 "data_size": 63488 00:13:11.611 } 00:13:11.611 ] 00:13:11.611 }' 00:13:11.611 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:11.611 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.129 90.50 IOPS, 271.50 MiB/s [2024-12-05T19:03:29.688Z] 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:12.129 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.129 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.129 [2024-12-05 19:03:29.456423] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:12.129 [2024-12-05 19:03:29.456455] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:12.129 00:13:12.130 Latency(us) 00:13:12.130 [2024-12-05T19:03:29.689Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:12.130 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:13:12.130 raid_bdev1 : 8.29 88.46 265.37 0.00 0.00 15951.20 262.93 116304.94 00:13:12.130 [2024-12-05T19:03:29.689Z] =================================================================================================================== 00:13:12.130 [2024-12-05T19:03:29.689Z] Total : 88.46 265.37 0.00 0.00 15951.20 262.93 116304.94 00:13:12.130 [2024-12-05 19:03:29.559287] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:12.130 [2024-12-05 19:03:29.559335] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:12.130 [2024-12-05 19:03:29.559426] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:12.130 [2024-12-05 19:03:29.559436] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:12.130 { 00:13:12.130 "results": [ 00:13:12.130 { 00:13:12.130 "job": "raid_bdev1", 00:13:12.130 "core_mask": "0x1", 00:13:12.130 "workload": "randrw", 00:13:12.130 "percentage": 50, 00:13:12.130 "status": "finished", 00:13:12.130 "queue_depth": 2, 00:13:12.130 "io_size": 3145728, 00:13:12.130 "runtime": 8.286521, 00:13:12.130 "iops": 88.45690489410454, 00:13:12.130 "mibps": 265.37071468231363, 00:13:12.130 "io_failed": 0, 00:13:12.130 "io_timeout": 0, 00:13:12.130 "avg_latency_us": 15951.20250212979, 00:13:12.130 "min_latency_us": 262.9310043668122, 00:13:12.130 "max_latency_us": 116304.93624454149 00:13:12.130 } 00:13:12.130 ], 00:13:12.130 "core_count": 1 00:13:12.130 } 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:12.130 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:13:12.390 /dev/nbd0 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:12.390 1+0 records in 00:13:12.390 1+0 records out 00:13:12.390 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000461376 s, 8.9 MB/s 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@728 -- # continue 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:12.390 19:03:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:13:12.651 /dev/nbd1 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:12.651 1+0 records in 00:13:12.651 1+0 records out 00:13:12.651 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000346655 s, 11.8 MB/s 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:12.651 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:12.911 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:13:13.171 /dev/nbd1 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # local i 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@877 -- # break 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:13.171 1+0 records in 00:13:13.171 1+0 records out 00:13:13.171 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000422125 s, 9.7 MB/s 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@890 -- # size=4096 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@893 -- # return 0 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:13.171 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:13.431 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:13.431 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:13.431 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:13.431 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:13.431 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:13.431 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:13.431 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:13.431 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:13.431 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:13.431 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:13.431 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:13.431 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:13.431 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:13.431 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:13.431 19:03:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.691 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.691 [2024-12-05 19:03:31.167789] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:13.691 [2024-12-05 19:03:31.167844] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:13.691 [2024-12-05 19:03:31.167867] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:13:13.691 [2024-12-05 19:03:31.167876] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:13.691 [2024-12-05 19:03:31.170083] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:13.691 spare 00:13:13.692 [2024-12-05 19:03:31.170183] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:13.692 [2024-12-05 19:03:31.170276] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:13.692 [2024-12-05 19:03:31.170321] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:13.692 [2024-12-05 19:03:31.170440] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:13.692 [2024-12-05 19:03:31.170518] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:13.692 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.692 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:13:13.692 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.692 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.952 [2024-12-05 19:03:31.270418] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:13:13.952 [2024-12-05 19:03:31.270439] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:13.952 [2024-12-05 19:03:31.270703] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000337b0 00:13:13.952 [2024-12-05 19:03:31.270843] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:13:13.952 [2024-12-05 19:03:31.270855] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:13:13.952 [2024-12-05 19:03:31.270970] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:13.952 "name": "raid_bdev1", 00:13:13.952 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:13.952 "strip_size_kb": 0, 00:13:13.952 "state": "online", 00:13:13.952 "raid_level": "raid1", 00:13:13.952 "superblock": true, 00:13:13.952 "num_base_bdevs": 4, 00:13:13.952 "num_base_bdevs_discovered": 3, 00:13:13.952 "num_base_bdevs_operational": 3, 00:13:13.952 "base_bdevs_list": [ 00:13:13.952 { 00:13:13.952 "name": "spare", 00:13:13.952 "uuid": "41cc6ffc-af87-5de8-a83b-08845eb57c3f", 00:13:13.952 "is_configured": true, 00:13:13.952 "data_offset": 2048, 00:13:13.952 "data_size": 63488 00:13:13.952 }, 00:13:13.952 { 00:13:13.952 "name": null, 00:13:13.952 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:13.952 "is_configured": false, 00:13:13.952 "data_offset": 2048, 00:13:13.952 "data_size": 63488 00:13:13.952 }, 00:13:13.952 { 00:13:13.952 "name": "BaseBdev3", 00:13:13.952 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:13.952 "is_configured": true, 00:13:13.952 "data_offset": 2048, 00:13:13.952 "data_size": 63488 00:13:13.952 }, 00:13:13.952 { 00:13:13.952 "name": "BaseBdev4", 00:13:13.952 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:13.952 "is_configured": true, 00:13:13.952 "data_offset": 2048, 00:13:13.952 "data_size": 63488 00:13:13.952 } 00:13:13.952 ] 00:13:13.952 }' 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:13.952 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.212 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:14.212 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:14.212 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:14.212 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:14.212 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:14.212 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.212 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:14.212 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.212 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.212 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.212 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:14.212 "name": "raid_bdev1", 00:13:14.212 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:14.212 "strip_size_kb": 0, 00:13:14.212 "state": "online", 00:13:14.212 "raid_level": "raid1", 00:13:14.212 "superblock": true, 00:13:14.213 "num_base_bdevs": 4, 00:13:14.213 "num_base_bdevs_discovered": 3, 00:13:14.213 "num_base_bdevs_operational": 3, 00:13:14.213 "base_bdevs_list": [ 00:13:14.213 { 00:13:14.213 "name": "spare", 00:13:14.213 "uuid": "41cc6ffc-af87-5de8-a83b-08845eb57c3f", 00:13:14.213 "is_configured": true, 00:13:14.213 "data_offset": 2048, 00:13:14.213 "data_size": 63488 00:13:14.213 }, 00:13:14.213 { 00:13:14.213 "name": null, 00:13:14.213 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.213 "is_configured": false, 00:13:14.213 "data_offset": 2048, 00:13:14.213 "data_size": 63488 00:13:14.213 }, 00:13:14.213 { 00:13:14.213 "name": "BaseBdev3", 00:13:14.213 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:14.213 "is_configured": true, 00:13:14.213 "data_offset": 2048, 00:13:14.213 "data_size": 63488 00:13:14.213 }, 00:13:14.213 { 00:13:14.213 "name": "BaseBdev4", 00:13:14.213 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:14.213 "is_configured": true, 00:13:14.213 "data_offset": 2048, 00:13:14.213 "data_size": 63488 00:13:14.213 } 00:13:14.213 ] 00:13:14.213 }' 00:13:14.213 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.473 [2024-12-05 19:03:31.890690] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:14.473 "name": "raid_bdev1", 00:13:14.473 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:14.473 "strip_size_kb": 0, 00:13:14.473 "state": "online", 00:13:14.473 "raid_level": "raid1", 00:13:14.473 "superblock": true, 00:13:14.473 "num_base_bdevs": 4, 00:13:14.473 "num_base_bdevs_discovered": 2, 00:13:14.473 "num_base_bdevs_operational": 2, 00:13:14.473 "base_bdevs_list": [ 00:13:14.473 { 00:13:14.473 "name": null, 00:13:14.473 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.473 "is_configured": false, 00:13:14.473 "data_offset": 0, 00:13:14.473 "data_size": 63488 00:13:14.473 }, 00:13:14.473 { 00:13:14.473 "name": null, 00:13:14.473 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.473 "is_configured": false, 00:13:14.473 "data_offset": 2048, 00:13:14.473 "data_size": 63488 00:13:14.473 }, 00:13:14.473 { 00:13:14.473 "name": "BaseBdev3", 00:13:14.473 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:14.473 "is_configured": true, 00:13:14.473 "data_offset": 2048, 00:13:14.473 "data_size": 63488 00:13:14.473 }, 00:13:14.473 { 00:13:14.473 "name": "BaseBdev4", 00:13:14.473 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:14.473 "is_configured": true, 00:13:14.473 "data_offset": 2048, 00:13:14.473 "data_size": 63488 00:13:14.473 } 00:13:14.473 ] 00:13:14.473 }' 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:14.473 19:03:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.043 19:03:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:15.043 19:03:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.043 19:03:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.043 [2024-12-05 19:03:32.345927] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:15.043 [2024-12-05 19:03:32.346132] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:15.043 [2024-12-05 19:03:32.346173] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:15.043 [2024-12-05 19:03:32.346213] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:15.043 [2024-12-05 19:03:32.350658] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000033880 00:13:15.043 19:03:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.043 19:03:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:13:15.043 [2024-12-05 19:03:32.352521] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:15.983 "name": "raid_bdev1", 00:13:15.983 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:15.983 "strip_size_kb": 0, 00:13:15.983 "state": "online", 00:13:15.983 "raid_level": "raid1", 00:13:15.983 "superblock": true, 00:13:15.983 "num_base_bdevs": 4, 00:13:15.983 "num_base_bdevs_discovered": 3, 00:13:15.983 "num_base_bdevs_operational": 3, 00:13:15.983 "process": { 00:13:15.983 "type": "rebuild", 00:13:15.983 "target": "spare", 00:13:15.983 "progress": { 00:13:15.983 "blocks": 20480, 00:13:15.983 "percent": 32 00:13:15.983 } 00:13:15.983 }, 00:13:15.983 "base_bdevs_list": [ 00:13:15.983 { 00:13:15.983 "name": "spare", 00:13:15.983 "uuid": "41cc6ffc-af87-5de8-a83b-08845eb57c3f", 00:13:15.983 "is_configured": true, 00:13:15.983 "data_offset": 2048, 00:13:15.983 "data_size": 63488 00:13:15.983 }, 00:13:15.983 { 00:13:15.983 "name": null, 00:13:15.983 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:15.983 "is_configured": false, 00:13:15.983 "data_offset": 2048, 00:13:15.983 "data_size": 63488 00:13:15.983 }, 00:13:15.983 { 00:13:15.983 "name": "BaseBdev3", 00:13:15.983 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:15.983 "is_configured": true, 00:13:15.983 "data_offset": 2048, 00:13:15.983 "data_size": 63488 00:13:15.983 }, 00:13:15.983 { 00:13:15.983 "name": "BaseBdev4", 00:13:15.983 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:15.983 "is_configured": true, 00:13:15.983 "data_offset": 2048, 00:13:15.983 "data_size": 63488 00:13:15.983 } 00:13:15.983 ] 00:13:15.983 }' 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.983 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:15.983 [2024-12-05 19:03:33.500652] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:16.243 [2024-12-05 19:03:33.556479] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:16.243 [2024-12-05 19:03:33.556591] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:16.243 [2024-12-05 19:03:33.556611] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:16.243 [2024-12-05 19:03:33.556618] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:16.243 "name": "raid_bdev1", 00:13:16.243 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:16.243 "strip_size_kb": 0, 00:13:16.243 "state": "online", 00:13:16.243 "raid_level": "raid1", 00:13:16.243 "superblock": true, 00:13:16.243 "num_base_bdevs": 4, 00:13:16.243 "num_base_bdevs_discovered": 2, 00:13:16.243 "num_base_bdevs_operational": 2, 00:13:16.243 "base_bdevs_list": [ 00:13:16.243 { 00:13:16.243 "name": null, 00:13:16.243 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:16.243 "is_configured": false, 00:13:16.243 "data_offset": 0, 00:13:16.243 "data_size": 63488 00:13:16.243 }, 00:13:16.243 { 00:13:16.243 "name": null, 00:13:16.243 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:16.243 "is_configured": false, 00:13:16.243 "data_offset": 2048, 00:13:16.243 "data_size": 63488 00:13:16.243 }, 00:13:16.243 { 00:13:16.243 "name": "BaseBdev3", 00:13:16.243 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:16.243 "is_configured": true, 00:13:16.243 "data_offset": 2048, 00:13:16.243 "data_size": 63488 00:13:16.243 }, 00:13:16.243 { 00:13:16.243 "name": "BaseBdev4", 00:13:16.243 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:16.243 "is_configured": true, 00:13:16.243 "data_offset": 2048, 00:13:16.243 "data_size": 63488 00:13:16.243 } 00:13:16.243 ] 00:13:16.243 }' 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:16.243 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.503 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:16.503 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:16.503 19:03:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:16.503 [2024-12-05 19:03:34.003947] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:16.503 [2024-12-05 19:03:34.004070] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:16.503 [2024-12-05 19:03:34.004115] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:13:16.503 [2024-12-05 19:03:34.004159] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:16.503 [2024-12-05 19:03:34.004588] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:16.503 [2024-12-05 19:03:34.004647] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:16.503 [2024-12-05 19:03:34.004782] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:16.503 [2024-12-05 19:03:34.004827] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:16.503 [2024-12-05 19:03:34.004866] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:16.503 [2024-12-05 19:03:34.004911] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:16.503 [2024-12-05 19:03:34.008955] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000033950 00:13:16.503 spare 00:13:16.503 19:03:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:16.503 19:03:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:13:16.503 [2024-12-05 19:03:34.010872] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:17.896 "name": "raid_bdev1", 00:13:17.896 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:17.896 "strip_size_kb": 0, 00:13:17.896 "state": "online", 00:13:17.896 "raid_level": "raid1", 00:13:17.896 "superblock": true, 00:13:17.896 "num_base_bdevs": 4, 00:13:17.896 "num_base_bdevs_discovered": 3, 00:13:17.896 "num_base_bdevs_operational": 3, 00:13:17.896 "process": { 00:13:17.896 "type": "rebuild", 00:13:17.896 "target": "spare", 00:13:17.896 "progress": { 00:13:17.896 "blocks": 20480, 00:13:17.896 "percent": 32 00:13:17.896 } 00:13:17.896 }, 00:13:17.896 "base_bdevs_list": [ 00:13:17.896 { 00:13:17.896 "name": "spare", 00:13:17.896 "uuid": "41cc6ffc-af87-5de8-a83b-08845eb57c3f", 00:13:17.896 "is_configured": true, 00:13:17.896 "data_offset": 2048, 00:13:17.896 "data_size": 63488 00:13:17.896 }, 00:13:17.896 { 00:13:17.896 "name": null, 00:13:17.896 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:17.896 "is_configured": false, 00:13:17.896 "data_offset": 2048, 00:13:17.896 "data_size": 63488 00:13:17.896 }, 00:13:17.896 { 00:13:17.896 "name": "BaseBdev3", 00:13:17.896 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:17.896 "is_configured": true, 00:13:17.896 "data_offset": 2048, 00:13:17.896 "data_size": 63488 00:13:17.896 }, 00:13:17.896 { 00:13:17.896 "name": "BaseBdev4", 00:13:17.896 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:17.896 "is_configured": true, 00:13:17.896 "data_offset": 2048, 00:13:17.896 "data_size": 63488 00:13:17.896 } 00:13:17.896 ] 00:13:17.896 }' 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:17.896 [2024-12-05 19:03:35.163455] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:17.896 [2024-12-05 19:03:35.214756] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:17.896 [2024-12-05 19:03:35.214810] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:17.896 [2024-12-05 19:03:35.214824] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:17.896 [2024-12-05 19:03:35.214832] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.896 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:17.897 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:17.897 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:17.897 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:17.897 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:17.897 "name": "raid_bdev1", 00:13:17.897 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:17.897 "strip_size_kb": 0, 00:13:17.897 "state": "online", 00:13:17.897 "raid_level": "raid1", 00:13:17.897 "superblock": true, 00:13:17.897 "num_base_bdevs": 4, 00:13:17.897 "num_base_bdevs_discovered": 2, 00:13:17.897 "num_base_bdevs_operational": 2, 00:13:17.897 "base_bdevs_list": [ 00:13:17.897 { 00:13:17.897 "name": null, 00:13:17.897 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:17.897 "is_configured": false, 00:13:17.897 "data_offset": 0, 00:13:17.897 "data_size": 63488 00:13:17.897 }, 00:13:17.897 { 00:13:17.897 "name": null, 00:13:17.897 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:17.897 "is_configured": false, 00:13:17.897 "data_offset": 2048, 00:13:17.897 "data_size": 63488 00:13:17.897 }, 00:13:17.897 { 00:13:17.897 "name": "BaseBdev3", 00:13:17.897 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:17.897 "is_configured": true, 00:13:17.897 "data_offset": 2048, 00:13:17.897 "data_size": 63488 00:13:17.897 }, 00:13:17.897 { 00:13:17.897 "name": "BaseBdev4", 00:13:17.897 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:17.897 "is_configured": true, 00:13:17.897 "data_offset": 2048, 00:13:17.897 "data_size": 63488 00:13:17.897 } 00:13:17.897 ] 00:13:17.897 }' 00:13:17.897 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:17.897 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.155 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:18.155 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:18.155 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:18.155 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:18.155 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:18.155 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:18.155 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:18.155 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:18.155 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.415 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:18.415 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:18.415 "name": "raid_bdev1", 00:13:18.415 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:18.415 "strip_size_kb": 0, 00:13:18.415 "state": "online", 00:13:18.415 "raid_level": "raid1", 00:13:18.415 "superblock": true, 00:13:18.415 "num_base_bdevs": 4, 00:13:18.415 "num_base_bdevs_discovered": 2, 00:13:18.415 "num_base_bdevs_operational": 2, 00:13:18.415 "base_bdevs_list": [ 00:13:18.415 { 00:13:18.415 "name": null, 00:13:18.415 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:18.415 "is_configured": false, 00:13:18.415 "data_offset": 0, 00:13:18.415 "data_size": 63488 00:13:18.415 }, 00:13:18.415 { 00:13:18.415 "name": null, 00:13:18.415 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:18.415 "is_configured": false, 00:13:18.415 "data_offset": 2048, 00:13:18.415 "data_size": 63488 00:13:18.415 }, 00:13:18.415 { 00:13:18.415 "name": "BaseBdev3", 00:13:18.415 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:18.415 "is_configured": true, 00:13:18.415 "data_offset": 2048, 00:13:18.415 "data_size": 63488 00:13:18.415 }, 00:13:18.415 { 00:13:18.415 "name": "BaseBdev4", 00:13:18.415 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:18.415 "is_configured": true, 00:13:18.415 "data_offset": 2048, 00:13:18.415 "data_size": 63488 00:13:18.415 } 00:13:18.415 ] 00:13:18.415 }' 00:13:18.415 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:18.415 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:18.415 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:18.415 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:18.415 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:13:18.415 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:18.415 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.415 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:18.415 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:18.415 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:18.415 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.415 [2024-12-05 19:03:35.826055] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:18.415 [2024-12-05 19:03:35.826112] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:18.415 [2024-12-05 19:03:35.826133] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:13:18.415 [2024-12-05 19:03:35.826145] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:18.415 [2024-12-05 19:03:35.826498] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:18.415 [2024-12-05 19:03:35.826517] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:18.415 [2024-12-05 19:03:35.826578] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:13:18.415 [2024-12-05 19:03:35.826592] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:18.415 [2024-12-05 19:03:35.826600] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:18.415 [2024-12-05 19:03:35.826616] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:13:18.415 BaseBdev1 00:13:18.415 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:18.415 19:03:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:13:19.354 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:19.354 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:19.354 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:19.354 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:19.354 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:19.355 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:19.355 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:19.355 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:19.355 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:19.355 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:19.355 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.355 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:19.355 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.355 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.355 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.355 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:19.355 "name": "raid_bdev1", 00:13:19.355 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:19.355 "strip_size_kb": 0, 00:13:19.355 "state": "online", 00:13:19.355 "raid_level": "raid1", 00:13:19.355 "superblock": true, 00:13:19.355 "num_base_bdevs": 4, 00:13:19.355 "num_base_bdevs_discovered": 2, 00:13:19.355 "num_base_bdevs_operational": 2, 00:13:19.355 "base_bdevs_list": [ 00:13:19.355 { 00:13:19.355 "name": null, 00:13:19.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:19.355 "is_configured": false, 00:13:19.355 "data_offset": 0, 00:13:19.355 "data_size": 63488 00:13:19.355 }, 00:13:19.355 { 00:13:19.355 "name": null, 00:13:19.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:19.355 "is_configured": false, 00:13:19.355 "data_offset": 2048, 00:13:19.355 "data_size": 63488 00:13:19.355 }, 00:13:19.355 { 00:13:19.355 "name": "BaseBdev3", 00:13:19.355 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:19.355 "is_configured": true, 00:13:19.355 "data_offset": 2048, 00:13:19.355 "data_size": 63488 00:13:19.355 }, 00:13:19.355 { 00:13:19.355 "name": "BaseBdev4", 00:13:19.355 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:19.355 "is_configured": true, 00:13:19.355 "data_offset": 2048, 00:13:19.355 "data_size": 63488 00:13:19.355 } 00:13:19.355 ] 00:13:19.355 }' 00:13:19.355 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:19.355 19:03:36 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.924 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:19.924 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:19.924 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:19.924 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:19.925 "name": "raid_bdev1", 00:13:19.925 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:19.925 "strip_size_kb": 0, 00:13:19.925 "state": "online", 00:13:19.925 "raid_level": "raid1", 00:13:19.925 "superblock": true, 00:13:19.925 "num_base_bdevs": 4, 00:13:19.925 "num_base_bdevs_discovered": 2, 00:13:19.925 "num_base_bdevs_operational": 2, 00:13:19.925 "base_bdevs_list": [ 00:13:19.925 { 00:13:19.925 "name": null, 00:13:19.925 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:19.925 "is_configured": false, 00:13:19.925 "data_offset": 0, 00:13:19.925 "data_size": 63488 00:13:19.925 }, 00:13:19.925 { 00:13:19.925 "name": null, 00:13:19.925 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:19.925 "is_configured": false, 00:13:19.925 "data_offset": 2048, 00:13:19.925 "data_size": 63488 00:13:19.925 }, 00:13:19.925 { 00:13:19.925 "name": "BaseBdev3", 00:13:19.925 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:19.925 "is_configured": true, 00:13:19.925 "data_offset": 2048, 00:13:19.925 "data_size": 63488 00:13:19.925 }, 00:13:19.925 { 00:13:19.925 "name": "BaseBdev4", 00:13:19.925 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:19.925 "is_configured": true, 00:13:19.925 "data_offset": 2048, 00:13:19.925 "data_size": 63488 00:13:19.925 } 00:13:19.925 ] 00:13:19.925 }' 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # local es=0 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:19.925 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.925 [2024-12-05 19:03:37.475441] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:19.925 [2024-12-05 19:03:37.475564] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:19.925 [2024-12-05 19:03:37.475576] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:19.925 request: 00:13:19.925 { 00:13:19.925 "base_bdev": "BaseBdev1", 00:13:19.925 "raid_bdev": "raid_bdev1", 00:13:19.925 "method": "bdev_raid_add_base_bdev", 00:13:19.925 "req_id": 1 00:13:19.925 } 00:13:19.925 Got JSON-RPC error response 00:13:19.925 response: 00:13:19.925 { 00:13:19.925 "code": -22, 00:13:19.925 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:13:19.925 } 00:13:20.185 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:13:20.185 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@655 -- # es=1 00:13:20.185 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:13:20.185 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:13:20.185 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:13:20.185 19:03:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:21.125 "name": "raid_bdev1", 00:13:21.125 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:21.125 "strip_size_kb": 0, 00:13:21.125 "state": "online", 00:13:21.125 "raid_level": "raid1", 00:13:21.125 "superblock": true, 00:13:21.125 "num_base_bdevs": 4, 00:13:21.125 "num_base_bdevs_discovered": 2, 00:13:21.125 "num_base_bdevs_operational": 2, 00:13:21.125 "base_bdevs_list": [ 00:13:21.125 { 00:13:21.125 "name": null, 00:13:21.125 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:21.125 "is_configured": false, 00:13:21.125 "data_offset": 0, 00:13:21.125 "data_size": 63488 00:13:21.125 }, 00:13:21.125 { 00:13:21.125 "name": null, 00:13:21.125 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:21.125 "is_configured": false, 00:13:21.125 "data_offset": 2048, 00:13:21.125 "data_size": 63488 00:13:21.125 }, 00:13:21.125 { 00:13:21.125 "name": "BaseBdev3", 00:13:21.125 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:21.125 "is_configured": true, 00:13:21.125 "data_offset": 2048, 00:13:21.125 "data_size": 63488 00:13:21.125 }, 00:13:21.125 { 00:13:21.125 "name": "BaseBdev4", 00:13:21.125 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:21.125 "is_configured": true, 00:13:21.125 "data_offset": 2048, 00:13:21.125 "data_size": 63488 00:13:21.125 } 00:13:21.125 ] 00:13:21.125 }' 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:21.125 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.385 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:21.385 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:21.385 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:21.385 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:21.385 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:21.385 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.386 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:21.386 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:21.386 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.646 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:21.646 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:21.646 "name": "raid_bdev1", 00:13:21.646 "uuid": "b454e770-82a3-4376-acbb-54177edb7a73", 00:13:21.646 "strip_size_kb": 0, 00:13:21.646 "state": "online", 00:13:21.646 "raid_level": "raid1", 00:13:21.646 "superblock": true, 00:13:21.646 "num_base_bdevs": 4, 00:13:21.646 "num_base_bdevs_discovered": 2, 00:13:21.646 "num_base_bdevs_operational": 2, 00:13:21.646 "base_bdevs_list": [ 00:13:21.646 { 00:13:21.646 "name": null, 00:13:21.646 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:21.646 "is_configured": false, 00:13:21.646 "data_offset": 0, 00:13:21.646 "data_size": 63488 00:13:21.646 }, 00:13:21.646 { 00:13:21.646 "name": null, 00:13:21.646 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:21.646 "is_configured": false, 00:13:21.646 "data_offset": 2048, 00:13:21.646 "data_size": 63488 00:13:21.646 }, 00:13:21.646 { 00:13:21.646 "name": "BaseBdev3", 00:13:21.646 "uuid": "772f7d95-d58e-5f86-9c34-87ee546a8ca5", 00:13:21.646 "is_configured": true, 00:13:21.646 "data_offset": 2048, 00:13:21.646 "data_size": 63488 00:13:21.646 }, 00:13:21.646 { 00:13:21.646 "name": "BaseBdev4", 00:13:21.646 "uuid": "1f4fe362-5b7e-51f2-8380-bcbc818af781", 00:13:21.646 "is_configured": true, 00:13:21.646 "data_offset": 2048, 00:13:21.646 "data_size": 63488 00:13:21.646 } 00:13:21.646 ] 00:13:21.646 }' 00:13:21.646 19:03:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:21.646 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:21.646 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:21.646 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:21.646 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 87820 00:13:21.646 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # '[' -z 87820 ']' 00:13:21.646 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@958 -- # kill -0 87820 00:13:21.646 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # uname 00:13:21.646 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:21.646 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 87820 00:13:21.646 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:21.646 killing process with pid 87820 00:13:21.646 Received shutdown signal, test time was about 17.861808 seconds 00:13:21.646 00:13:21.646 Latency(us) 00:13:21.646 [2024-12-05T19:03:39.205Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:21.646 [2024-12-05T19:03:39.205Z] =================================================================================================================== 00:13:21.646 [2024-12-05T19:03:39.205Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:21.646 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:21.646 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@972 -- # echo 'killing process with pid 87820' 00:13:21.646 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@973 -- # kill 87820 00:13:21.646 [2024-12-05 19:03:39.113019] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:21.646 [2024-12-05 19:03:39.113126] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:21.646 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@978 -- # wait 87820 00:13:21.646 [2024-12-05 19:03:39.113187] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:21.646 [2024-12-05 19:03:39.113196] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:13:21.646 [2024-12-05 19:03:39.157615] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:21.907 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:13:21.907 ************************************ 00:13:21.907 END TEST raid_rebuild_test_sb_io 00:13:21.907 ************************************ 00:13:21.907 00:13:21.907 real 0m19.807s 00:13:21.907 user 0m26.414s 00:13:21.907 sys 0m2.618s 00:13:21.907 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:21.907 19:03:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.907 19:03:39 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:13:21.907 19:03:39 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 3 false 00:13:21.907 19:03:39 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:13:21.907 19:03:39 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:21.907 19:03:39 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:21.907 ************************************ 00:13:21.907 START TEST raid5f_state_function_test 00:13:21.907 ************************************ 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 3 false 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=88412 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 88412' 00:13:21.907 Process raid pid: 88412 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 88412 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 88412 ']' 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:21.907 19:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:21.907 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:22.168 19:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:22.168 19:03:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:22.168 [2024-12-05 19:03:39.544129] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:13:22.168 [2024-12-05 19:03:39.544341] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:22.168 [2024-12-05 19:03:39.700367] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:22.168 [2024-12-05 19:03:39.725763] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:22.428 [2024-12-05 19:03:39.769245] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:22.428 [2024-12-05 19:03:39.769357] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.021 [2024-12-05 19:03:40.364644] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:23.021 [2024-12-05 19:03:40.364768] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:23.021 [2024-12-05 19:03:40.364811] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:23.021 [2024-12-05 19:03:40.364836] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:23.021 [2024-12-05 19:03:40.364854] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:23.021 [2024-12-05 19:03:40.364887] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:23.021 "name": "Existed_Raid", 00:13:23.021 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.021 "strip_size_kb": 64, 00:13:23.021 "state": "configuring", 00:13:23.021 "raid_level": "raid5f", 00:13:23.021 "superblock": false, 00:13:23.021 "num_base_bdevs": 3, 00:13:23.021 "num_base_bdevs_discovered": 0, 00:13:23.021 "num_base_bdevs_operational": 3, 00:13:23.021 "base_bdevs_list": [ 00:13:23.021 { 00:13:23.021 "name": "BaseBdev1", 00:13:23.021 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.021 "is_configured": false, 00:13:23.021 "data_offset": 0, 00:13:23.021 "data_size": 0 00:13:23.021 }, 00:13:23.021 { 00:13:23.021 "name": "BaseBdev2", 00:13:23.021 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.021 "is_configured": false, 00:13:23.021 "data_offset": 0, 00:13:23.021 "data_size": 0 00:13:23.021 }, 00:13:23.021 { 00:13:23.021 "name": "BaseBdev3", 00:13:23.021 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.021 "is_configured": false, 00:13:23.021 "data_offset": 0, 00:13:23.021 "data_size": 0 00:13:23.021 } 00:13:23.021 ] 00:13:23.021 }' 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:23.021 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.280 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:23.280 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.280 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.540 [2024-12-05 19:03:40.843759] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:23.540 [2024-12-05 19:03:40.843835] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.540 [2024-12-05 19:03:40.855774] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:23.540 [2024-12-05 19:03:40.855852] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:23.540 [2024-12-05 19:03:40.855878] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:23.540 [2024-12-05 19:03:40.855900] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:23.540 [2024-12-05 19:03:40.855917] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:23.540 [2024-12-05 19:03:40.855937] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.540 [2024-12-05 19:03:40.876700] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:23.540 BaseBdev1 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.540 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.540 [ 00:13:23.540 { 00:13:23.540 "name": "BaseBdev1", 00:13:23.540 "aliases": [ 00:13:23.540 "a73853a9-331f-4c10-a485-0ab06728ab80" 00:13:23.540 ], 00:13:23.540 "product_name": "Malloc disk", 00:13:23.540 "block_size": 512, 00:13:23.540 "num_blocks": 65536, 00:13:23.540 "uuid": "a73853a9-331f-4c10-a485-0ab06728ab80", 00:13:23.540 "assigned_rate_limits": { 00:13:23.540 "rw_ios_per_sec": 0, 00:13:23.540 "rw_mbytes_per_sec": 0, 00:13:23.540 "r_mbytes_per_sec": 0, 00:13:23.540 "w_mbytes_per_sec": 0 00:13:23.540 }, 00:13:23.540 "claimed": true, 00:13:23.540 "claim_type": "exclusive_write", 00:13:23.540 "zoned": false, 00:13:23.540 "supported_io_types": { 00:13:23.540 "read": true, 00:13:23.540 "write": true, 00:13:23.540 "unmap": true, 00:13:23.540 "flush": true, 00:13:23.540 "reset": true, 00:13:23.540 "nvme_admin": false, 00:13:23.540 "nvme_io": false, 00:13:23.540 "nvme_io_md": false, 00:13:23.540 "write_zeroes": true, 00:13:23.540 "zcopy": true, 00:13:23.540 "get_zone_info": false, 00:13:23.540 "zone_management": false, 00:13:23.541 "zone_append": false, 00:13:23.541 "compare": false, 00:13:23.541 "compare_and_write": false, 00:13:23.541 "abort": true, 00:13:23.541 "seek_hole": false, 00:13:23.541 "seek_data": false, 00:13:23.541 "copy": true, 00:13:23.541 "nvme_iov_md": false 00:13:23.541 }, 00:13:23.541 "memory_domains": [ 00:13:23.541 { 00:13:23.541 "dma_device_id": "system", 00:13:23.541 "dma_device_type": 1 00:13:23.541 }, 00:13:23.541 { 00:13:23.541 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:23.541 "dma_device_type": 2 00:13:23.541 } 00:13:23.541 ], 00:13:23.541 "driver_specific": {} 00:13:23.541 } 00:13:23.541 ] 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:23.541 "name": "Existed_Raid", 00:13:23.541 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.541 "strip_size_kb": 64, 00:13:23.541 "state": "configuring", 00:13:23.541 "raid_level": "raid5f", 00:13:23.541 "superblock": false, 00:13:23.541 "num_base_bdevs": 3, 00:13:23.541 "num_base_bdevs_discovered": 1, 00:13:23.541 "num_base_bdevs_operational": 3, 00:13:23.541 "base_bdevs_list": [ 00:13:23.541 { 00:13:23.541 "name": "BaseBdev1", 00:13:23.541 "uuid": "a73853a9-331f-4c10-a485-0ab06728ab80", 00:13:23.541 "is_configured": true, 00:13:23.541 "data_offset": 0, 00:13:23.541 "data_size": 65536 00:13:23.541 }, 00:13:23.541 { 00:13:23.541 "name": "BaseBdev2", 00:13:23.541 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.541 "is_configured": false, 00:13:23.541 "data_offset": 0, 00:13:23.541 "data_size": 0 00:13:23.541 }, 00:13:23.541 { 00:13:23.541 "name": "BaseBdev3", 00:13:23.541 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.541 "is_configured": false, 00:13:23.541 "data_offset": 0, 00:13:23.541 "data_size": 0 00:13:23.541 } 00:13:23.541 ] 00:13:23.541 }' 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:23.541 19:03:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.109 [2024-12-05 19:03:41.367879] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:24.109 [2024-12-05 19:03:41.367973] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.109 [2024-12-05 19:03:41.379894] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:24.109 [2024-12-05 19:03:41.381719] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:24.109 [2024-12-05 19:03:41.381754] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:24.109 [2024-12-05 19:03:41.381763] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:24.109 [2024-12-05 19:03:41.381774] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:24.109 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:24.110 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.110 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:24.110 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.110 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.110 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.110 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:24.110 "name": "Existed_Raid", 00:13:24.110 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:24.110 "strip_size_kb": 64, 00:13:24.110 "state": "configuring", 00:13:24.110 "raid_level": "raid5f", 00:13:24.110 "superblock": false, 00:13:24.110 "num_base_bdevs": 3, 00:13:24.110 "num_base_bdevs_discovered": 1, 00:13:24.110 "num_base_bdevs_operational": 3, 00:13:24.110 "base_bdevs_list": [ 00:13:24.110 { 00:13:24.110 "name": "BaseBdev1", 00:13:24.110 "uuid": "a73853a9-331f-4c10-a485-0ab06728ab80", 00:13:24.110 "is_configured": true, 00:13:24.110 "data_offset": 0, 00:13:24.110 "data_size": 65536 00:13:24.110 }, 00:13:24.110 { 00:13:24.110 "name": "BaseBdev2", 00:13:24.110 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:24.110 "is_configured": false, 00:13:24.110 "data_offset": 0, 00:13:24.110 "data_size": 0 00:13:24.110 }, 00:13:24.110 { 00:13:24.110 "name": "BaseBdev3", 00:13:24.110 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:24.110 "is_configured": false, 00:13:24.110 "data_offset": 0, 00:13:24.110 "data_size": 0 00:13:24.110 } 00:13:24.110 ] 00:13:24.110 }' 00:13:24.110 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:24.110 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.370 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:24.370 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.370 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.370 [2024-12-05 19:03:41.878101] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:24.370 BaseBdev2 00:13:24.370 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.370 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:13:24.370 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:24.370 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:24.370 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.371 [ 00:13:24.371 { 00:13:24.371 "name": "BaseBdev2", 00:13:24.371 "aliases": [ 00:13:24.371 "343267ee-93cc-43dc-b766-2400e5192961" 00:13:24.371 ], 00:13:24.371 "product_name": "Malloc disk", 00:13:24.371 "block_size": 512, 00:13:24.371 "num_blocks": 65536, 00:13:24.371 "uuid": "343267ee-93cc-43dc-b766-2400e5192961", 00:13:24.371 "assigned_rate_limits": { 00:13:24.371 "rw_ios_per_sec": 0, 00:13:24.371 "rw_mbytes_per_sec": 0, 00:13:24.371 "r_mbytes_per_sec": 0, 00:13:24.371 "w_mbytes_per_sec": 0 00:13:24.371 }, 00:13:24.371 "claimed": true, 00:13:24.371 "claim_type": "exclusive_write", 00:13:24.371 "zoned": false, 00:13:24.371 "supported_io_types": { 00:13:24.371 "read": true, 00:13:24.371 "write": true, 00:13:24.371 "unmap": true, 00:13:24.371 "flush": true, 00:13:24.371 "reset": true, 00:13:24.371 "nvme_admin": false, 00:13:24.371 "nvme_io": false, 00:13:24.371 "nvme_io_md": false, 00:13:24.371 "write_zeroes": true, 00:13:24.371 "zcopy": true, 00:13:24.371 "get_zone_info": false, 00:13:24.371 "zone_management": false, 00:13:24.371 "zone_append": false, 00:13:24.371 "compare": false, 00:13:24.371 "compare_and_write": false, 00:13:24.371 "abort": true, 00:13:24.371 "seek_hole": false, 00:13:24.371 "seek_data": false, 00:13:24.371 "copy": true, 00:13:24.371 "nvme_iov_md": false 00:13:24.371 }, 00:13:24.371 "memory_domains": [ 00:13:24.371 { 00:13:24.371 "dma_device_id": "system", 00:13:24.371 "dma_device_type": 1 00:13:24.371 }, 00:13:24.371 { 00:13:24.371 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:24.371 "dma_device_type": 2 00:13:24.371 } 00:13:24.371 ], 00:13:24.371 "driver_specific": {} 00:13:24.371 } 00:13:24.371 ] 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.371 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.631 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.631 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:24.631 "name": "Existed_Raid", 00:13:24.631 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:24.631 "strip_size_kb": 64, 00:13:24.631 "state": "configuring", 00:13:24.631 "raid_level": "raid5f", 00:13:24.631 "superblock": false, 00:13:24.631 "num_base_bdevs": 3, 00:13:24.631 "num_base_bdevs_discovered": 2, 00:13:24.631 "num_base_bdevs_operational": 3, 00:13:24.631 "base_bdevs_list": [ 00:13:24.631 { 00:13:24.631 "name": "BaseBdev1", 00:13:24.631 "uuid": "a73853a9-331f-4c10-a485-0ab06728ab80", 00:13:24.631 "is_configured": true, 00:13:24.631 "data_offset": 0, 00:13:24.631 "data_size": 65536 00:13:24.631 }, 00:13:24.631 { 00:13:24.631 "name": "BaseBdev2", 00:13:24.631 "uuid": "343267ee-93cc-43dc-b766-2400e5192961", 00:13:24.631 "is_configured": true, 00:13:24.631 "data_offset": 0, 00:13:24.631 "data_size": 65536 00:13:24.631 }, 00:13:24.631 { 00:13:24.631 "name": "BaseBdev3", 00:13:24.631 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:24.631 "is_configured": false, 00:13:24.631 "data_offset": 0, 00:13:24.631 "data_size": 0 00:13:24.631 } 00:13:24.631 ] 00:13:24.631 }' 00:13:24.631 19:03:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:24.631 19:03:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.890 [2024-12-05 19:03:42.400346] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:24.890 [2024-12-05 19:03:42.400528] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:13:24.890 [2024-12-05 19:03:42.400577] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:24.890 [2024-12-05 19:03:42.401558] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:13:24.890 [2024-12-05 19:03:42.403255] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:13:24.890 [2024-12-05 19:03:42.403302] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:13:24.890 BaseBdev3 00:13:24.890 [2024-12-05 19:03:42.404111] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:24.890 [ 00:13:24.890 { 00:13:24.890 "name": "BaseBdev3", 00:13:24.890 "aliases": [ 00:13:24.890 "94f5792c-f748-416d-aede-0444d64c3dca" 00:13:24.890 ], 00:13:24.890 "product_name": "Malloc disk", 00:13:24.890 "block_size": 512, 00:13:24.890 "num_blocks": 65536, 00:13:24.890 "uuid": "94f5792c-f748-416d-aede-0444d64c3dca", 00:13:24.890 "assigned_rate_limits": { 00:13:24.890 "rw_ios_per_sec": 0, 00:13:24.890 "rw_mbytes_per_sec": 0, 00:13:24.890 "r_mbytes_per_sec": 0, 00:13:24.890 "w_mbytes_per_sec": 0 00:13:24.890 }, 00:13:24.890 "claimed": true, 00:13:24.890 "claim_type": "exclusive_write", 00:13:24.890 "zoned": false, 00:13:24.890 "supported_io_types": { 00:13:24.890 "read": true, 00:13:24.890 "write": true, 00:13:24.890 "unmap": true, 00:13:24.890 "flush": true, 00:13:24.890 "reset": true, 00:13:24.890 "nvme_admin": false, 00:13:24.890 "nvme_io": false, 00:13:24.890 "nvme_io_md": false, 00:13:24.890 "write_zeroes": true, 00:13:24.890 "zcopy": true, 00:13:24.890 "get_zone_info": false, 00:13:24.890 "zone_management": false, 00:13:24.890 "zone_append": false, 00:13:24.890 "compare": false, 00:13:24.890 "compare_and_write": false, 00:13:24.890 "abort": true, 00:13:24.890 "seek_hole": false, 00:13:24.890 "seek_data": false, 00:13:24.890 "copy": true, 00:13:24.890 "nvme_iov_md": false 00:13:24.890 }, 00:13:24.890 "memory_domains": [ 00:13:24.890 { 00:13:24.890 "dma_device_id": "system", 00:13:24.890 "dma_device_type": 1 00:13:24.890 }, 00:13:24.890 { 00:13:24.890 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:24.890 "dma_device_type": 2 00:13:24.890 } 00:13:24.890 ], 00:13:24.890 "driver_specific": {} 00:13:24.890 } 00:13:24.890 ] 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:24.890 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:25.149 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.149 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:25.149 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.149 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.149 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.149 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:25.149 "name": "Existed_Raid", 00:13:25.149 "uuid": "1c66905c-966c-4e95-a7be-698bb445fba1", 00:13:25.149 "strip_size_kb": 64, 00:13:25.149 "state": "online", 00:13:25.149 "raid_level": "raid5f", 00:13:25.150 "superblock": false, 00:13:25.150 "num_base_bdevs": 3, 00:13:25.150 "num_base_bdevs_discovered": 3, 00:13:25.150 "num_base_bdevs_operational": 3, 00:13:25.150 "base_bdevs_list": [ 00:13:25.150 { 00:13:25.150 "name": "BaseBdev1", 00:13:25.150 "uuid": "a73853a9-331f-4c10-a485-0ab06728ab80", 00:13:25.150 "is_configured": true, 00:13:25.150 "data_offset": 0, 00:13:25.150 "data_size": 65536 00:13:25.150 }, 00:13:25.150 { 00:13:25.150 "name": "BaseBdev2", 00:13:25.150 "uuid": "343267ee-93cc-43dc-b766-2400e5192961", 00:13:25.150 "is_configured": true, 00:13:25.150 "data_offset": 0, 00:13:25.150 "data_size": 65536 00:13:25.150 }, 00:13:25.150 { 00:13:25.150 "name": "BaseBdev3", 00:13:25.150 "uuid": "94f5792c-f748-416d-aede-0444d64c3dca", 00:13:25.150 "is_configured": true, 00:13:25.150 "data_offset": 0, 00:13:25.150 "data_size": 65536 00:13:25.150 } 00:13:25.150 ] 00:13:25.150 }' 00:13:25.150 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:25.150 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.409 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:13:25.409 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:25.409 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:25.409 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:25.409 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:25.409 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:25.409 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:25.409 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:25.409 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.409 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.409 [2024-12-05 19:03:42.919847] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:25.409 19:03:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.409 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:25.409 "name": "Existed_Raid", 00:13:25.409 "aliases": [ 00:13:25.409 "1c66905c-966c-4e95-a7be-698bb445fba1" 00:13:25.409 ], 00:13:25.409 "product_name": "Raid Volume", 00:13:25.409 "block_size": 512, 00:13:25.409 "num_blocks": 131072, 00:13:25.409 "uuid": "1c66905c-966c-4e95-a7be-698bb445fba1", 00:13:25.409 "assigned_rate_limits": { 00:13:25.409 "rw_ios_per_sec": 0, 00:13:25.409 "rw_mbytes_per_sec": 0, 00:13:25.409 "r_mbytes_per_sec": 0, 00:13:25.409 "w_mbytes_per_sec": 0 00:13:25.409 }, 00:13:25.409 "claimed": false, 00:13:25.409 "zoned": false, 00:13:25.409 "supported_io_types": { 00:13:25.409 "read": true, 00:13:25.409 "write": true, 00:13:25.409 "unmap": false, 00:13:25.409 "flush": false, 00:13:25.409 "reset": true, 00:13:25.409 "nvme_admin": false, 00:13:25.409 "nvme_io": false, 00:13:25.409 "nvme_io_md": false, 00:13:25.409 "write_zeroes": true, 00:13:25.409 "zcopy": false, 00:13:25.409 "get_zone_info": false, 00:13:25.409 "zone_management": false, 00:13:25.409 "zone_append": false, 00:13:25.409 "compare": false, 00:13:25.409 "compare_and_write": false, 00:13:25.409 "abort": false, 00:13:25.409 "seek_hole": false, 00:13:25.409 "seek_data": false, 00:13:25.409 "copy": false, 00:13:25.409 "nvme_iov_md": false 00:13:25.409 }, 00:13:25.409 "driver_specific": { 00:13:25.409 "raid": { 00:13:25.409 "uuid": "1c66905c-966c-4e95-a7be-698bb445fba1", 00:13:25.409 "strip_size_kb": 64, 00:13:25.409 "state": "online", 00:13:25.409 "raid_level": "raid5f", 00:13:25.409 "superblock": false, 00:13:25.409 "num_base_bdevs": 3, 00:13:25.409 "num_base_bdevs_discovered": 3, 00:13:25.409 "num_base_bdevs_operational": 3, 00:13:25.409 "base_bdevs_list": [ 00:13:25.409 { 00:13:25.409 "name": "BaseBdev1", 00:13:25.409 "uuid": "a73853a9-331f-4c10-a485-0ab06728ab80", 00:13:25.409 "is_configured": true, 00:13:25.409 "data_offset": 0, 00:13:25.409 "data_size": 65536 00:13:25.409 }, 00:13:25.409 { 00:13:25.409 "name": "BaseBdev2", 00:13:25.409 "uuid": "343267ee-93cc-43dc-b766-2400e5192961", 00:13:25.409 "is_configured": true, 00:13:25.409 "data_offset": 0, 00:13:25.409 "data_size": 65536 00:13:25.409 }, 00:13:25.409 { 00:13:25.409 "name": "BaseBdev3", 00:13:25.409 "uuid": "94f5792c-f748-416d-aede-0444d64c3dca", 00:13:25.409 "is_configured": true, 00:13:25.409 "data_offset": 0, 00:13:25.409 "data_size": 65536 00:13:25.409 } 00:13:25.409 ] 00:13:25.409 } 00:13:25.409 } 00:13:25.409 }' 00:13:25.409 19:03:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:13:25.668 BaseBdev2 00:13:25.668 BaseBdev3' 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.668 [2024-12-05 19:03:43.207193] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:25.668 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:25.928 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.928 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:25.928 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:25.928 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.928 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:25.928 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:25.928 "name": "Existed_Raid", 00:13:25.928 "uuid": "1c66905c-966c-4e95-a7be-698bb445fba1", 00:13:25.928 "strip_size_kb": 64, 00:13:25.928 "state": "online", 00:13:25.928 "raid_level": "raid5f", 00:13:25.928 "superblock": false, 00:13:25.928 "num_base_bdevs": 3, 00:13:25.928 "num_base_bdevs_discovered": 2, 00:13:25.928 "num_base_bdevs_operational": 2, 00:13:25.928 "base_bdevs_list": [ 00:13:25.928 { 00:13:25.928 "name": null, 00:13:25.928 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:25.928 "is_configured": false, 00:13:25.928 "data_offset": 0, 00:13:25.928 "data_size": 65536 00:13:25.928 }, 00:13:25.928 { 00:13:25.928 "name": "BaseBdev2", 00:13:25.928 "uuid": "343267ee-93cc-43dc-b766-2400e5192961", 00:13:25.928 "is_configured": true, 00:13:25.928 "data_offset": 0, 00:13:25.928 "data_size": 65536 00:13:25.928 }, 00:13:25.928 { 00:13:25.928 "name": "BaseBdev3", 00:13:25.928 "uuid": "94f5792c-f748-416d-aede-0444d64c3dca", 00:13:25.928 "is_configured": true, 00:13:25.928 "data_offset": 0, 00:13:25.928 "data_size": 65536 00:13:25.928 } 00:13:25.928 ] 00:13:25.928 }' 00:13:25.928 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:25.928 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.187 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:13:26.187 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:26.187 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.187 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.187 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:26.187 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.187 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.448 [2024-12-05 19:03:43.765411] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:26.448 [2024-12-05 19:03:43.765562] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:26.448 [2024-12-05 19:03:43.776747] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.448 [2024-12-05 19:03:43.836682] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:26.448 [2024-12-05 19:03:43.836739] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.448 BaseBdev2 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.448 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.449 [ 00:13:26.449 { 00:13:26.449 "name": "BaseBdev2", 00:13:26.449 "aliases": [ 00:13:26.449 "70a8126b-4c1c-4eed-86b3-f99e71711a1b" 00:13:26.449 ], 00:13:26.449 "product_name": "Malloc disk", 00:13:26.449 "block_size": 512, 00:13:26.449 "num_blocks": 65536, 00:13:26.449 "uuid": "70a8126b-4c1c-4eed-86b3-f99e71711a1b", 00:13:26.449 "assigned_rate_limits": { 00:13:26.449 "rw_ios_per_sec": 0, 00:13:26.449 "rw_mbytes_per_sec": 0, 00:13:26.449 "r_mbytes_per_sec": 0, 00:13:26.449 "w_mbytes_per_sec": 0 00:13:26.449 }, 00:13:26.449 "claimed": false, 00:13:26.449 "zoned": false, 00:13:26.449 "supported_io_types": { 00:13:26.449 "read": true, 00:13:26.449 "write": true, 00:13:26.449 "unmap": true, 00:13:26.449 "flush": true, 00:13:26.449 "reset": true, 00:13:26.449 "nvme_admin": false, 00:13:26.449 "nvme_io": false, 00:13:26.449 "nvme_io_md": false, 00:13:26.449 "write_zeroes": true, 00:13:26.449 "zcopy": true, 00:13:26.449 "get_zone_info": false, 00:13:26.449 "zone_management": false, 00:13:26.449 "zone_append": false, 00:13:26.449 "compare": false, 00:13:26.449 "compare_and_write": false, 00:13:26.449 "abort": true, 00:13:26.449 "seek_hole": false, 00:13:26.449 "seek_data": false, 00:13:26.449 "copy": true, 00:13:26.449 "nvme_iov_md": false 00:13:26.449 }, 00:13:26.449 "memory_domains": [ 00:13:26.449 { 00:13:26.449 "dma_device_id": "system", 00:13:26.449 "dma_device_type": 1 00:13:26.449 }, 00:13:26.449 { 00:13:26.449 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:26.449 "dma_device_type": 2 00:13:26.449 } 00:13:26.449 ], 00:13:26.449 "driver_specific": {} 00:13:26.449 } 00:13:26.449 ] 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.449 BaseBdev3 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.449 19:03:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.449 [ 00:13:26.449 { 00:13:26.449 "name": "BaseBdev3", 00:13:26.449 "aliases": [ 00:13:26.449 "d747d90c-e672-4015-9f99-f22970a40248" 00:13:26.449 ], 00:13:26.449 "product_name": "Malloc disk", 00:13:26.449 "block_size": 512, 00:13:26.449 "num_blocks": 65536, 00:13:26.449 "uuid": "d747d90c-e672-4015-9f99-f22970a40248", 00:13:26.449 "assigned_rate_limits": { 00:13:26.449 "rw_ios_per_sec": 0, 00:13:26.449 "rw_mbytes_per_sec": 0, 00:13:26.449 "r_mbytes_per_sec": 0, 00:13:26.449 "w_mbytes_per_sec": 0 00:13:26.449 }, 00:13:26.449 "claimed": false, 00:13:26.449 "zoned": false, 00:13:26.449 "supported_io_types": { 00:13:26.449 "read": true, 00:13:26.449 "write": true, 00:13:26.449 "unmap": true, 00:13:26.449 "flush": true, 00:13:26.449 "reset": true, 00:13:26.449 "nvme_admin": false, 00:13:26.449 "nvme_io": false, 00:13:26.449 "nvme_io_md": false, 00:13:26.449 "write_zeroes": true, 00:13:26.449 "zcopy": true, 00:13:26.449 "get_zone_info": false, 00:13:26.449 "zone_management": false, 00:13:26.449 "zone_append": false, 00:13:26.449 "compare": false, 00:13:26.449 "compare_and_write": false, 00:13:26.449 "abort": true, 00:13:26.449 "seek_hole": false, 00:13:26.449 "seek_data": false, 00:13:26.450 "copy": true, 00:13:26.450 "nvme_iov_md": false 00:13:26.450 }, 00:13:26.450 "memory_domains": [ 00:13:26.450 { 00:13:26.450 "dma_device_id": "system", 00:13:26.450 "dma_device_type": 1 00:13:26.450 }, 00:13:26.450 { 00:13:26.450 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:26.450 "dma_device_type": 2 00:13:26.450 } 00:13:26.450 ], 00:13:26.450 "driver_specific": {} 00:13:26.450 } 00:13:26.450 ] 00:13:26.450 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.450 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:26.450 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:26.450 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:26.708 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:26.708 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.708 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.708 [2024-12-05 19:03:44.010540] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:26.708 [2024-12-05 19:03:44.010675] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:26.708 [2024-12-05 19:03:44.010716] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:26.708 [2024-12-05 19:03:44.012515] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:26.708 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.708 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:26.708 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:26.708 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:26.708 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:26.708 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:26.708 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:26.708 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:26.708 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:26.709 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:26.709 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:26.709 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.709 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.709 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.709 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:26.709 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.709 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:26.709 "name": "Existed_Raid", 00:13:26.709 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:26.709 "strip_size_kb": 64, 00:13:26.709 "state": "configuring", 00:13:26.709 "raid_level": "raid5f", 00:13:26.709 "superblock": false, 00:13:26.709 "num_base_bdevs": 3, 00:13:26.709 "num_base_bdevs_discovered": 2, 00:13:26.709 "num_base_bdevs_operational": 3, 00:13:26.709 "base_bdevs_list": [ 00:13:26.709 { 00:13:26.709 "name": "BaseBdev1", 00:13:26.709 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:26.709 "is_configured": false, 00:13:26.709 "data_offset": 0, 00:13:26.709 "data_size": 0 00:13:26.709 }, 00:13:26.709 { 00:13:26.709 "name": "BaseBdev2", 00:13:26.709 "uuid": "70a8126b-4c1c-4eed-86b3-f99e71711a1b", 00:13:26.709 "is_configured": true, 00:13:26.709 "data_offset": 0, 00:13:26.709 "data_size": 65536 00:13:26.709 }, 00:13:26.709 { 00:13:26.709 "name": "BaseBdev3", 00:13:26.709 "uuid": "d747d90c-e672-4015-9f99-f22970a40248", 00:13:26.709 "is_configured": true, 00:13:26.709 "data_offset": 0, 00:13:26.709 "data_size": 65536 00:13:26.709 } 00:13:26.709 ] 00:13:26.709 }' 00:13:26.709 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:26.709 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.968 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:26.968 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.968 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.968 [2024-12-05 19:03:44.481704] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:26.968 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:26.968 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:26.968 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:26.968 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:26.968 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:26.968 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:26.968 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:26.968 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:26.968 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:26.969 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:26.969 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:26.969 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.969 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.969 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.969 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:26.969 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.229 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:27.229 "name": "Existed_Raid", 00:13:27.229 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:27.229 "strip_size_kb": 64, 00:13:27.229 "state": "configuring", 00:13:27.229 "raid_level": "raid5f", 00:13:27.229 "superblock": false, 00:13:27.229 "num_base_bdevs": 3, 00:13:27.229 "num_base_bdevs_discovered": 1, 00:13:27.229 "num_base_bdevs_operational": 3, 00:13:27.229 "base_bdevs_list": [ 00:13:27.229 { 00:13:27.229 "name": "BaseBdev1", 00:13:27.229 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:27.229 "is_configured": false, 00:13:27.229 "data_offset": 0, 00:13:27.229 "data_size": 0 00:13:27.229 }, 00:13:27.229 { 00:13:27.229 "name": null, 00:13:27.229 "uuid": "70a8126b-4c1c-4eed-86b3-f99e71711a1b", 00:13:27.229 "is_configured": false, 00:13:27.229 "data_offset": 0, 00:13:27.229 "data_size": 65536 00:13:27.229 }, 00:13:27.229 { 00:13:27.229 "name": "BaseBdev3", 00:13:27.229 "uuid": "d747d90c-e672-4015-9f99-f22970a40248", 00:13:27.229 "is_configured": true, 00:13:27.229 "data_offset": 0, 00:13:27.229 "data_size": 65536 00:13:27.229 } 00:13:27.229 ] 00:13:27.229 }' 00:13:27.229 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:27.229 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.490 [2024-12-05 19:03:44.975780] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:27.490 BaseBdev1 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.490 19:03:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.490 [ 00:13:27.490 { 00:13:27.490 "name": "BaseBdev1", 00:13:27.490 "aliases": [ 00:13:27.490 "e26eec0b-39c8-4891-aba4-e3d5a48cd7ce" 00:13:27.490 ], 00:13:27.490 "product_name": "Malloc disk", 00:13:27.490 "block_size": 512, 00:13:27.490 "num_blocks": 65536, 00:13:27.490 "uuid": "e26eec0b-39c8-4891-aba4-e3d5a48cd7ce", 00:13:27.490 "assigned_rate_limits": { 00:13:27.490 "rw_ios_per_sec": 0, 00:13:27.490 "rw_mbytes_per_sec": 0, 00:13:27.490 "r_mbytes_per_sec": 0, 00:13:27.490 "w_mbytes_per_sec": 0 00:13:27.490 }, 00:13:27.490 "claimed": true, 00:13:27.490 "claim_type": "exclusive_write", 00:13:27.490 "zoned": false, 00:13:27.490 "supported_io_types": { 00:13:27.490 "read": true, 00:13:27.490 "write": true, 00:13:27.490 "unmap": true, 00:13:27.490 "flush": true, 00:13:27.490 "reset": true, 00:13:27.491 "nvme_admin": false, 00:13:27.491 "nvme_io": false, 00:13:27.491 "nvme_io_md": false, 00:13:27.491 "write_zeroes": true, 00:13:27.491 "zcopy": true, 00:13:27.491 "get_zone_info": false, 00:13:27.491 "zone_management": false, 00:13:27.491 "zone_append": false, 00:13:27.491 "compare": false, 00:13:27.491 "compare_and_write": false, 00:13:27.491 "abort": true, 00:13:27.491 "seek_hole": false, 00:13:27.491 "seek_data": false, 00:13:27.491 "copy": true, 00:13:27.491 "nvme_iov_md": false 00:13:27.491 }, 00:13:27.491 "memory_domains": [ 00:13:27.491 { 00:13:27.491 "dma_device_id": "system", 00:13:27.491 "dma_device_type": 1 00:13:27.491 }, 00:13:27.491 { 00:13:27.491 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:27.491 "dma_device_type": 2 00:13:27.491 } 00:13:27.491 ], 00:13:27.491 "driver_specific": {} 00:13:27.491 } 00:13:27.491 ] 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.491 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.750 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:27.750 "name": "Existed_Raid", 00:13:27.750 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:27.750 "strip_size_kb": 64, 00:13:27.750 "state": "configuring", 00:13:27.750 "raid_level": "raid5f", 00:13:27.750 "superblock": false, 00:13:27.750 "num_base_bdevs": 3, 00:13:27.750 "num_base_bdevs_discovered": 2, 00:13:27.750 "num_base_bdevs_operational": 3, 00:13:27.750 "base_bdevs_list": [ 00:13:27.750 { 00:13:27.750 "name": "BaseBdev1", 00:13:27.750 "uuid": "e26eec0b-39c8-4891-aba4-e3d5a48cd7ce", 00:13:27.750 "is_configured": true, 00:13:27.750 "data_offset": 0, 00:13:27.750 "data_size": 65536 00:13:27.750 }, 00:13:27.750 { 00:13:27.750 "name": null, 00:13:27.750 "uuid": "70a8126b-4c1c-4eed-86b3-f99e71711a1b", 00:13:27.750 "is_configured": false, 00:13:27.750 "data_offset": 0, 00:13:27.750 "data_size": 65536 00:13:27.750 }, 00:13:27.750 { 00:13:27.750 "name": "BaseBdev3", 00:13:27.750 "uuid": "d747d90c-e672-4015-9f99-f22970a40248", 00:13:27.750 "is_configured": true, 00:13:27.750 "data_offset": 0, 00:13:27.750 "data_size": 65536 00:13:27.750 } 00:13:27.750 ] 00:13:27.750 }' 00:13:27.750 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:27.750 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.008 [2024-12-05 19:03:45.514904] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.008 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.267 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:28.267 "name": "Existed_Raid", 00:13:28.267 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:28.267 "strip_size_kb": 64, 00:13:28.267 "state": "configuring", 00:13:28.267 "raid_level": "raid5f", 00:13:28.267 "superblock": false, 00:13:28.267 "num_base_bdevs": 3, 00:13:28.267 "num_base_bdevs_discovered": 1, 00:13:28.267 "num_base_bdevs_operational": 3, 00:13:28.267 "base_bdevs_list": [ 00:13:28.267 { 00:13:28.267 "name": "BaseBdev1", 00:13:28.267 "uuid": "e26eec0b-39c8-4891-aba4-e3d5a48cd7ce", 00:13:28.267 "is_configured": true, 00:13:28.267 "data_offset": 0, 00:13:28.267 "data_size": 65536 00:13:28.267 }, 00:13:28.267 { 00:13:28.267 "name": null, 00:13:28.267 "uuid": "70a8126b-4c1c-4eed-86b3-f99e71711a1b", 00:13:28.267 "is_configured": false, 00:13:28.267 "data_offset": 0, 00:13:28.267 "data_size": 65536 00:13:28.267 }, 00:13:28.267 { 00:13:28.267 "name": null, 00:13:28.267 "uuid": "d747d90c-e672-4015-9f99-f22970a40248", 00:13:28.267 "is_configured": false, 00:13:28.267 "data_offset": 0, 00:13:28.267 "data_size": 65536 00:13:28.267 } 00:13:28.267 ] 00:13:28.267 }' 00:13:28.267 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:28.267 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.527 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:28.527 19:03:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.527 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.527 19:03:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.527 [2024-12-05 19:03:46.018070] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:28.527 "name": "Existed_Raid", 00:13:28.527 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:28.527 "strip_size_kb": 64, 00:13:28.527 "state": "configuring", 00:13:28.527 "raid_level": "raid5f", 00:13:28.527 "superblock": false, 00:13:28.527 "num_base_bdevs": 3, 00:13:28.527 "num_base_bdevs_discovered": 2, 00:13:28.527 "num_base_bdevs_operational": 3, 00:13:28.527 "base_bdevs_list": [ 00:13:28.527 { 00:13:28.527 "name": "BaseBdev1", 00:13:28.527 "uuid": "e26eec0b-39c8-4891-aba4-e3d5a48cd7ce", 00:13:28.527 "is_configured": true, 00:13:28.527 "data_offset": 0, 00:13:28.527 "data_size": 65536 00:13:28.527 }, 00:13:28.527 { 00:13:28.527 "name": null, 00:13:28.527 "uuid": "70a8126b-4c1c-4eed-86b3-f99e71711a1b", 00:13:28.527 "is_configured": false, 00:13:28.527 "data_offset": 0, 00:13:28.527 "data_size": 65536 00:13:28.527 }, 00:13:28.527 { 00:13:28.527 "name": "BaseBdev3", 00:13:28.527 "uuid": "d747d90c-e672-4015-9f99-f22970a40248", 00:13:28.527 "is_configured": true, 00:13:28.527 "data_offset": 0, 00:13:28.527 "data_size": 65536 00:13:28.527 } 00:13:28.527 ] 00:13:28.527 }' 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:28.527 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.098 [2024-12-05 19:03:46.437346] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:29.098 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:29.099 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:29.099 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.099 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.099 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.099 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:29.099 "name": "Existed_Raid", 00:13:29.099 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:29.099 "strip_size_kb": 64, 00:13:29.099 "state": "configuring", 00:13:29.099 "raid_level": "raid5f", 00:13:29.099 "superblock": false, 00:13:29.099 "num_base_bdevs": 3, 00:13:29.099 "num_base_bdevs_discovered": 1, 00:13:29.099 "num_base_bdevs_operational": 3, 00:13:29.099 "base_bdevs_list": [ 00:13:29.099 { 00:13:29.099 "name": null, 00:13:29.099 "uuid": "e26eec0b-39c8-4891-aba4-e3d5a48cd7ce", 00:13:29.099 "is_configured": false, 00:13:29.099 "data_offset": 0, 00:13:29.099 "data_size": 65536 00:13:29.099 }, 00:13:29.099 { 00:13:29.099 "name": null, 00:13:29.099 "uuid": "70a8126b-4c1c-4eed-86b3-f99e71711a1b", 00:13:29.099 "is_configured": false, 00:13:29.099 "data_offset": 0, 00:13:29.099 "data_size": 65536 00:13:29.099 }, 00:13:29.099 { 00:13:29.099 "name": "BaseBdev3", 00:13:29.099 "uuid": "d747d90c-e672-4015-9f99-f22970a40248", 00:13:29.099 "is_configured": true, 00:13:29.099 "data_offset": 0, 00:13:29.099 "data_size": 65536 00:13:29.099 } 00:13:29.099 ] 00:13:29.099 }' 00:13:29.099 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:29.099 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.371 [2024-12-05 19:03:46.899025] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.371 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.649 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.649 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:29.649 "name": "Existed_Raid", 00:13:29.649 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:29.649 "strip_size_kb": 64, 00:13:29.649 "state": "configuring", 00:13:29.649 "raid_level": "raid5f", 00:13:29.649 "superblock": false, 00:13:29.649 "num_base_bdevs": 3, 00:13:29.649 "num_base_bdevs_discovered": 2, 00:13:29.649 "num_base_bdevs_operational": 3, 00:13:29.649 "base_bdevs_list": [ 00:13:29.649 { 00:13:29.649 "name": null, 00:13:29.649 "uuid": "e26eec0b-39c8-4891-aba4-e3d5a48cd7ce", 00:13:29.649 "is_configured": false, 00:13:29.649 "data_offset": 0, 00:13:29.649 "data_size": 65536 00:13:29.649 }, 00:13:29.649 { 00:13:29.649 "name": "BaseBdev2", 00:13:29.649 "uuid": "70a8126b-4c1c-4eed-86b3-f99e71711a1b", 00:13:29.649 "is_configured": true, 00:13:29.649 "data_offset": 0, 00:13:29.649 "data_size": 65536 00:13:29.649 }, 00:13:29.649 { 00:13:29.649 "name": "BaseBdev3", 00:13:29.649 "uuid": "d747d90c-e672-4015-9f99-f22970a40248", 00:13:29.649 "is_configured": true, 00:13:29.649 "data_offset": 0, 00:13:29.649 "data_size": 65536 00:13:29.649 } 00:13:29.649 ] 00:13:29.649 }' 00:13:29.649 19:03:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:29.649 19:03:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u e26eec0b-39c8-4891-aba4-e3d5a48cd7ce 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.918 [2024-12-05 19:03:47.448642] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:13:29.918 [2024-12-05 19:03:47.448700] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:13:29.918 [2024-12-05 19:03:47.448710] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:29.918 [2024-12-05 19:03:47.448976] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:13:29.918 [2024-12-05 19:03:47.449369] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:13:29.918 [2024-12-05 19:03:47.449385] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:13:29.918 [2024-12-05 19:03:47.449558] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:29.918 NewBaseBdev 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:29.918 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.918 [ 00:13:30.183 { 00:13:30.183 "name": "NewBaseBdev", 00:13:30.183 "aliases": [ 00:13:30.183 "e26eec0b-39c8-4891-aba4-e3d5a48cd7ce" 00:13:30.183 ], 00:13:30.183 "product_name": "Malloc disk", 00:13:30.183 "block_size": 512, 00:13:30.183 "num_blocks": 65536, 00:13:30.183 "uuid": "e26eec0b-39c8-4891-aba4-e3d5a48cd7ce", 00:13:30.183 "assigned_rate_limits": { 00:13:30.183 "rw_ios_per_sec": 0, 00:13:30.183 "rw_mbytes_per_sec": 0, 00:13:30.183 "r_mbytes_per_sec": 0, 00:13:30.183 "w_mbytes_per_sec": 0 00:13:30.183 }, 00:13:30.183 "claimed": true, 00:13:30.183 "claim_type": "exclusive_write", 00:13:30.183 "zoned": false, 00:13:30.183 "supported_io_types": { 00:13:30.183 "read": true, 00:13:30.183 "write": true, 00:13:30.183 "unmap": true, 00:13:30.183 "flush": true, 00:13:30.183 "reset": true, 00:13:30.183 "nvme_admin": false, 00:13:30.183 "nvme_io": false, 00:13:30.183 "nvme_io_md": false, 00:13:30.183 "write_zeroes": true, 00:13:30.183 "zcopy": true, 00:13:30.183 "get_zone_info": false, 00:13:30.183 "zone_management": false, 00:13:30.183 "zone_append": false, 00:13:30.183 "compare": false, 00:13:30.183 "compare_and_write": false, 00:13:30.183 "abort": true, 00:13:30.183 "seek_hole": false, 00:13:30.183 "seek_data": false, 00:13:30.183 "copy": true, 00:13:30.183 "nvme_iov_md": false 00:13:30.183 }, 00:13:30.183 "memory_domains": [ 00:13:30.183 { 00:13:30.183 "dma_device_id": "system", 00:13:30.183 "dma_device_type": 1 00:13:30.183 }, 00:13:30.183 { 00:13:30.183 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:30.183 "dma_device_type": 2 00:13:30.183 } 00:13:30.183 ], 00:13:30.183 "driver_specific": {} 00:13:30.183 } 00:13:30.183 ] 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:30.183 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:30.183 "name": "Existed_Raid", 00:13:30.183 "uuid": "ef4d0e26-5ead-495d-a045-0428f8582980", 00:13:30.183 "strip_size_kb": 64, 00:13:30.183 "state": "online", 00:13:30.183 "raid_level": "raid5f", 00:13:30.183 "superblock": false, 00:13:30.183 "num_base_bdevs": 3, 00:13:30.183 "num_base_bdevs_discovered": 3, 00:13:30.183 "num_base_bdevs_operational": 3, 00:13:30.183 "base_bdevs_list": [ 00:13:30.183 { 00:13:30.183 "name": "NewBaseBdev", 00:13:30.183 "uuid": "e26eec0b-39c8-4891-aba4-e3d5a48cd7ce", 00:13:30.183 "is_configured": true, 00:13:30.183 "data_offset": 0, 00:13:30.183 "data_size": 65536 00:13:30.183 }, 00:13:30.183 { 00:13:30.183 "name": "BaseBdev2", 00:13:30.183 "uuid": "70a8126b-4c1c-4eed-86b3-f99e71711a1b", 00:13:30.183 "is_configured": true, 00:13:30.183 "data_offset": 0, 00:13:30.183 "data_size": 65536 00:13:30.184 }, 00:13:30.184 { 00:13:30.184 "name": "BaseBdev3", 00:13:30.184 "uuid": "d747d90c-e672-4015-9f99-f22970a40248", 00:13:30.184 "is_configured": true, 00:13:30.184 "data_offset": 0, 00:13:30.184 "data_size": 65536 00:13:30.184 } 00:13:30.184 ] 00:13:30.184 }' 00:13:30.184 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:30.184 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.444 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:13:30.444 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:30.444 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:30.444 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:30.444 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:30.444 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:30.444 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:30.444 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:30.444 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.444 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:30.444 [2024-12-05 19:03:47.928125] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:30.444 19:03:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:30.444 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:30.444 "name": "Existed_Raid", 00:13:30.444 "aliases": [ 00:13:30.444 "ef4d0e26-5ead-495d-a045-0428f8582980" 00:13:30.444 ], 00:13:30.444 "product_name": "Raid Volume", 00:13:30.444 "block_size": 512, 00:13:30.444 "num_blocks": 131072, 00:13:30.444 "uuid": "ef4d0e26-5ead-495d-a045-0428f8582980", 00:13:30.444 "assigned_rate_limits": { 00:13:30.444 "rw_ios_per_sec": 0, 00:13:30.444 "rw_mbytes_per_sec": 0, 00:13:30.444 "r_mbytes_per_sec": 0, 00:13:30.444 "w_mbytes_per_sec": 0 00:13:30.444 }, 00:13:30.444 "claimed": false, 00:13:30.444 "zoned": false, 00:13:30.444 "supported_io_types": { 00:13:30.444 "read": true, 00:13:30.444 "write": true, 00:13:30.444 "unmap": false, 00:13:30.444 "flush": false, 00:13:30.444 "reset": true, 00:13:30.444 "nvme_admin": false, 00:13:30.444 "nvme_io": false, 00:13:30.444 "nvme_io_md": false, 00:13:30.444 "write_zeroes": true, 00:13:30.444 "zcopy": false, 00:13:30.444 "get_zone_info": false, 00:13:30.444 "zone_management": false, 00:13:30.444 "zone_append": false, 00:13:30.444 "compare": false, 00:13:30.444 "compare_and_write": false, 00:13:30.444 "abort": false, 00:13:30.444 "seek_hole": false, 00:13:30.444 "seek_data": false, 00:13:30.444 "copy": false, 00:13:30.444 "nvme_iov_md": false 00:13:30.444 }, 00:13:30.444 "driver_specific": { 00:13:30.444 "raid": { 00:13:30.444 "uuid": "ef4d0e26-5ead-495d-a045-0428f8582980", 00:13:30.444 "strip_size_kb": 64, 00:13:30.444 "state": "online", 00:13:30.444 "raid_level": "raid5f", 00:13:30.444 "superblock": false, 00:13:30.444 "num_base_bdevs": 3, 00:13:30.444 "num_base_bdevs_discovered": 3, 00:13:30.444 "num_base_bdevs_operational": 3, 00:13:30.444 "base_bdevs_list": [ 00:13:30.444 { 00:13:30.444 "name": "NewBaseBdev", 00:13:30.444 "uuid": "e26eec0b-39c8-4891-aba4-e3d5a48cd7ce", 00:13:30.444 "is_configured": true, 00:13:30.444 "data_offset": 0, 00:13:30.444 "data_size": 65536 00:13:30.444 }, 00:13:30.444 { 00:13:30.444 "name": "BaseBdev2", 00:13:30.444 "uuid": "70a8126b-4c1c-4eed-86b3-f99e71711a1b", 00:13:30.444 "is_configured": true, 00:13:30.444 "data_offset": 0, 00:13:30.444 "data_size": 65536 00:13:30.444 }, 00:13:30.444 { 00:13:30.444 "name": "BaseBdev3", 00:13:30.444 "uuid": "d747d90c-e672-4015-9f99-f22970a40248", 00:13:30.444 "is_configured": true, 00:13:30.444 "data_offset": 0, 00:13:30.444 "data_size": 65536 00:13:30.444 } 00:13:30.444 ] 00:13:30.444 } 00:13:30.444 } 00:13:30.444 }' 00:13:30.444 19:03:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:30.704 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:13:30.704 BaseBdev2 00:13:30.704 BaseBdev3' 00:13:30.704 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:30.704 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:30.704 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:30.704 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:13:30.704 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:30.704 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:30.704 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.704 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:30.704 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:30.704 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:30.704 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:30.704 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.705 [2024-12-05 19:03:48.227403] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:30.705 [2024-12-05 19:03:48.227473] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:30.705 [2024-12-05 19:03:48.227554] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:30.705 [2024-12-05 19:03:48.227825] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:30.705 [2024-12-05 19:03:48.227882] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 88412 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 88412 ']' 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # kill -0 88412 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # uname 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:30.705 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 88412 00:13:30.964 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:30.964 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:30.964 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 88412' 00:13:30.964 killing process with pid 88412 00:13:30.964 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@973 -- # kill 88412 00:13:30.964 [2024-12-05 19:03:48.274020] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:30.964 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@978 -- # wait 88412 00:13:30.964 [2024-12-05 19:03:48.304265] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:30.964 19:03:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:13:30.964 00:13:30.964 real 0m9.074s 00:13:30.964 user 0m15.437s 00:13:30.964 sys 0m2.024s 00:13:30.964 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:30.964 19:03:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.964 ************************************ 00:13:30.964 END TEST raid5f_state_function_test 00:13:30.964 ************************************ 00:13:31.225 19:03:48 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 3 true 00:13:31.225 19:03:48 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:13:31.225 19:03:48 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:31.225 19:03:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:31.225 ************************************ 00:13:31.225 START TEST raid5f_state_function_test_sb 00:13:31.225 ************************************ 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 3 true 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=88963 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 88963' 00:13:31.225 Process raid pid: 88963 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 88963 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 88963 ']' 00:13:31.225 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:31.225 19:03:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:31.225 [2024-12-05 19:03:48.696659] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:13:31.225 [2024-12-05 19:03:48.696796] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:31.486 [2024-12-05 19:03:48.851120] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:31.486 [2024-12-05 19:03:48.876124] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:31.486 [2024-12-05 19:03:48.919343] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:31.486 [2024-12-05 19:03:48.919379] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:32.057 [2024-12-05 19:03:49.506831] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:32.057 [2024-12-05 19:03:49.506898] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:32.057 [2024-12-05 19:03:49.506908] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:32.057 [2024-12-05 19:03:49.506918] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:32.057 [2024-12-05 19:03:49.506924] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:32.057 [2024-12-05 19:03:49.506937] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:32.057 "name": "Existed_Raid", 00:13:32.057 "uuid": "0f2c3ccb-3bb9-4a1c-bf12-b623b93f4c95", 00:13:32.057 "strip_size_kb": 64, 00:13:32.057 "state": "configuring", 00:13:32.057 "raid_level": "raid5f", 00:13:32.057 "superblock": true, 00:13:32.057 "num_base_bdevs": 3, 00:13:32.057 "num_base_bdevs_discovered": 0, 00:13:32.057 "num_base_bdevs_operational": 3, 00:13:32.057 "base_bdevs_list": [ 00:13:32.057 { 00:13:32.057 "name": "BaseBdev1", 00:13:32.057 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.057 "is_configured": false, 00:13:32.057 "data_offset": 0, 00:13:32.057 "data_size": 0 00:13:32.057 }, 00:13:32.057 { 00:13:32.057 "name": "BaseBdev2", 00:13:32.057 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.057 "is_configured": false, 00:13:32.057 "data_offset": 0, 00:13:32.057 "data_size": 0 00:13:32.057 }, 00:13:32.057 { 00:13:32.057 "name": "BaseBdev3", 00:13:32.057 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.057 "is_configured": false, 00:13:32.057 "data_offset": 0, 00:13:32.057 "data_size": 0 00:13:32.057 } 00:13:32.057 ] 00:13:32.057 }' 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:32.057 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:32.628 [2024-12-05 19:03:49.945959] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:32.628 [2024-12-05 19:03:49.946053] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:32.628 [2024-12-05 19:03:49.957972] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:32.628 [2024-12-05 19:03:49.958014] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:32.628 [2024-12-05 19:03:49.958022] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:32.628 [2024-12-05 19:03:49.958031] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:32.628 [2024-12-05 19:03:49.958037] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:32.628 [2024-12-05 19:03:49.958046] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:32.628 [2024-12-05 19:03:49.978856] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:32.628 BaseBdev1 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.628 19:03:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:32.628 [ 00:13:32.628 { 00:13:32.628 "name": "BaseBdev1", 00:13:32.628 "aliases": [ 00:13:32.628 "0a55f15a-0deb-44dc-8c17-3f0219c32465" 00:13:32.628 ], 00:13:32.628 "product_name": "Malloc disk", 00:13:32.628 "block_size": 512, 00:13:32.628 "num_blocks": 65536, 00:13:32.628 "uuid": "0a55f15a-0deb-44dc-8c17-3f0219c32465", 00:13:32.628 "assigned_rate_limits": { 00:13:32.628 "rw_ios_per_sec": 0, 00:13:32.628 "rw_mbytes_per_sec": 0, 00:13:32.628 "r_mbytes_per_sec": 0, 00:13:32.628 "w_mbytes_per_sec": 0 00:13:32.628 }, 00:13:32.628 "claimed": true, 00:13:32.628 "claim_type": "exclusive_write", 00:13:32.628 "zoned": false, 00:13:32.628 "supported_io_types": { 00:13:32.628 "read": true, 00:13:32.628 "write": true, 00:13:32.628 "unmap": true, 00:13:32.628 "flush": true, 00:13:32.628 "reset": true, 00:13:32.628 "nvme_admin": false, 00:13:32.628 "nvme_io": false, 00:13:32.628 "nvme_io_md": false, 00:13:32.628 "write_zeroes": true, 00:13:32.628 "zcopy": true, 00:13:32.628 "get_zone_info": false, 00:13:32.628 "zone_management": false, 00:13:32.628 "zone_append": false, 00:13:32.628 "compare": false, 00:13:32.628 "compare_and_write": false, 00:13:32.628 "abort": true, 00:13:32.628 "seek_hole": false, 00:13:32.628 "seek_data": false, 00:13:32.628 "copy": true, 00:13:32.628 "nvme_iov_md": false 00:13:32.628 }, 00:13:32.628 "memory_domains": [ 00:13:32.628 { 00:13:32.628 "dma_device_id": "system", 00:13:32.628 "dma_device_type": 1 00:13:32.628 }, 00:13:32.628 { 00:13:32.628 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:32.628 "dma_device_type": 2 00:13:32.628 } 00:13:32.628 ], 00:13:32.628 "driver_specific": {} 00:13:32.628 } 00:13:32.628 ] 00:13:32.628 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.628 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:32.628 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:32.628 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:32.628 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:32.628 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:32.628 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:32.628 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:32.629 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:32.629 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:32.629 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:32.629 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:32.629 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.629 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:32.629 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.629 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:32.629 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:32.629 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:32.629 "name": "Existed_Raid", 00:13:32.629 "uuid": "8316bfcc-655f-43f9-8bd3-426b30ae6fea", 00:13:32.629 "strip_size_kb": 64, 00:13:32.629 "state": "configuring", 00:13:32.629 "raid_level": "raid5f", 00:13:32.629 "superblock": true, 00:13:32.629 "num_base_bdevs": 3, 00:13:32.629 "num_base_bdevs_discovered": 1, 00:13:32.629 "num_base_bdevs_operational": 3, 00:13:32.629 "base_bdevs_list": [ 00:13:32.629 { 00:13:32.629 "name": "BaseBdev1", 00:13:32.629 "uuid": "0a55f15a-0deb-44dc-8c17-3f0219c32465", 00:13:32.629 "is_configured": true, 00:13:32.629 "data_offset": 2048, 00:13:32.629 "data_size": 63488 00:13:32.629 }, 00:13:32.629 { 00:13:32.629 "name": "BaseBdev2", 00:13:32.629 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.629 "is_configured": false, 00:13:32.629 "data_offset": 0, 00:13:32.629 "data_size": 0 00:13:32.629 }, 00:13:32.629 { 00:13:32.629 "name": "BaseBdev3", 00:13:32.629 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.629 "is_configured": false, 00:13:32.629 "data_offset": 0, 00:13:32.629 "data_size": 0 00:13:32.629 } 00:13:32.629 ] 00:13:32.629 }' 00:13:32.629 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:32.629 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:32.889 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:32.889 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:32.889 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:33.150 [2024-12-05 19:03:50.450063] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:33.150 [2024-12-05 19:03:50.450103] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:33.150 [2024-12-05 19:03:50.462077] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:33.150 [2024-12-05 19:03:50.463906] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:33.150 [2024-12-05 19:03:50.463946] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:33.150 [2024-12-05 19:03:50.463964] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:33.150 [2024-12-05 19:03:50.463975] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:33.150 "name": "Existed_Raid", 00:13:33.150 "uuid": "4eb50b6c-05c6-4ede-a8a8-92ee536525a6", 00:13:33.150 "strip_size_kb": 64, 00:13:33.150 "state": "configuring", 00:13:33.150 "raid_level": "raid5f", 00:13:33.150 "superblock": true, 00:13:33.150 "num_base_bdevs": 3, 00:13:33.150 "num_base_bdevs_discovered": 1, 00:13:33.150 "num_base_bdevs_operational": 3, 00:13:33.150 "base_bdevs_list": [ 00:13:33.150 { 00:13:33.150 "name": "BaseBdev1", 00:13:33.150 "uuid": "0a55f15a-0deb-44dc-8c17-3f0219c32465", 00:13:33.150 "is_configured": true, 00:13:33.150 "data_offset": 2048, 00:13:33.150 "data_size": 63488 00:13:33.150 }, 00:13:33.150 { 00:13:33.150 "name": "BaseBdev2", 00:13:33.150 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.150 "is_configured": false, 00:13:33.150 "data_offset": 0, 00:13:33.150 "data_size": 0 00:13:33.150 }, 00:13:33.150 { 00:13:33.150 "name": "BaseBdev3", 00:13:33.150 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.150 "is_configured": false, 00:13:33.150 "data_offset": 0, 00:13:33.150 "data_size": 0 00:13:33.150 } 00:13:33.150 ] 00:13:33.150 }' 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:33.150 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:33.410 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:33.410 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.410 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:33.410 [2024-12-05 19:03:50.948353] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:33.410 BaseBdev2 00:13:33.410 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.410 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:13:33.410 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:33.410 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:33.410 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:33.410 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:33.410 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:33.410 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:33.410 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.411 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:33.411 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.411 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:33.411 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.411 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:33.669 [ 00:13:33.669 { 00:13:33.669 "name": "BaseBdev2", 00:13:33.669 "aliases": [ 00:13:33.669 "4f6530a5-a639-4dcb-a98d-e9dfa95e7bb0" 00:13:33.669 ], 00:13:33.670 "product_name": "Malloc disk", 00:13:33.670 "block_size": 512, 00:13:33.670 "num_blocks": 65536, 00:13:33.670 "uuid": "4f6530a5-a639-4dcb-a98d-e9dfa95e7bb0", 00:13:33.670 "assigned_rate_limits": { 00:13:33.670 "rw_ios_per_sec": 0, 00:13:33.670 "rw_mbytes_per_sec": 0, 00:13:33.670 "r_mbytes_per_sec": 0, 00:13:33.670 "w_mbytes_per_sec": 0 00:13:33.670 }, 00:13:33.670 "claimed": true, 00:13:33.670 "claim_type": "exclusive_write", 00:13:33.670 "zoned": false, 00:13:33.670 "supported_io_types": { 00:13:33.670 "read": true, 00:13:33.670 "write": true, 00:13:33.670 "unmap": true, 00:13:33.670 "flush": true, 00:13:33.670 "reset": true, 00:13:33.670 "nvme_admin": false, 00:13:33.670 "nvme_io": false, 00:13:33.670 "nvme_io_md": false, 00:13:33.670 "write_zeroes": true, 00:13:33.670 "zcopy": true, 00:13:33.670 "get_zone_info": false, 00:13:33.670 "zone_management": false, 00:13:33.670 "zone_append": false, 00:13:33.670 "compare": false, 00:13:33.670 "compare_and_write": false, 00:13:33.670 "abort": true, 00:13:33.670 "seek_hole": false, 00:13:33.670 "seek_data": false, 00:13:33.670 "copy": true, 00:13:33.670 "nvme_iov_md": false 00:13:33.670 }, 00:13:33.670 "memory_domains": [ 00:13:33.670 { 00:13:33.670 "dma_device_id": "system", 00:13:33.670 "dma_device_type": 1 00:13:33.670 }, 00:13:33.670 { 00:13:33.670 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:33.670 "dma_device_type": 2 00:13:33.670 } 00:13:33.670 ], 00:13:33.670 "driver_specific": {} 00:13:33.670 } 00:13:33.670 ] 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.670 19:03:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:33.670 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.670 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:33.670 "name": "Existed_Raid", 00:13:33.670 "uuid": "4eb50b6c-05c6-4ede-a8a8-92ee536525a6", 00:13:33.670 "strip_size_kb": 64, 00:13:33.670 "state": "configuring", 00:13:33.670 "raid_level": "raid5f", 00:13:33.670 "superblock": true, 00:13:33.670 "num_base_bdevs": 3, 00:13:33.670 "num_base_bdevs_discovered": 2, 00:13:33.670 "num_base_bdevs_operational": 3, 00:13:33.670 "base_bdevs_list": [ 00:13:33.670 { 00:13:33.670 "name": "BaseBdev1", 00:13:33.670 "uuid": "0a55f15a-0deb-44dc-8c17-3f0219c32465", 00:13:33.670 "is_configured": true, 00:13:33.670 "data_offset": 2048, 00:13:33.670 "data_size": 63488 00:13:33.670 }, 00:13:33.670 { 00:13:33.670 "name": "BaseBdev2", 00:13:33.670 "uuid": "4f6530a5-a639-4dcb-a98d-e9dfa95e7bb0", 00:13:33.670 "is_configured": true, 00:13:33.670 "data_offset": 2048, 00:13:33.670 "data_size": 63488 00:13:33.670 }, 00:13:33.670 { 00:13:33.670 "name": "BaseBdev3", 00:13:33.670 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.670 "is_configured": false, 00:13:33.670 "data_offset": 0, 00:13:33.670 "data_size": 0 00:13:33.670 } 00:13:33.670 ] 00:13:33.670 }' 00:13:33.670 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:33.670 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:33.929 [2024-12-05 19:03:51.424255] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:33.929 [2024-12-05 19:03:51.424889] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:13:33.929 [2024-12-05 19:03:51.424957] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:33.929 BaseBdev3 00:13:33.929 [2024-12-05 19:03:51.425951] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:33.929 [2024-12-05 19:03:51.427644] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:13:33.929 [2024-12-05 19:03:51.427747] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:33.929 [2024-12-05 19:03:51.428278] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:33.929 [ 00:13:33.929 { 00:13:33.929 "name": "BaseBdev3", 00:13:33.929 "aliases": [ 00:13:33.929 "97d988b6-4370-4582-84a9-714d52885e84" 00:13:33.929 ], 00:13:33.929 "product_name": "Malloc disk", 00:13:33.929 "block_size": 512, 00:13:33.929 "num_blocks": 65536, 00:13:33.929 "uuid": "97d988b6-4370-4582-84a9-714d52885e84", 00:13:33.929 "assigned_rate_limits": { 00:13:33.929 "rw_ios_per_sec": 0, 00:13:33.929 "rw_mbytes_per_sec": 0, 00:13:33.929 "r_mbytes_per_sec": 0, 00:13:33.929 "w_mbytes_per_sec": 0 00:13:33.929 }, 00:13:33.929 "claimed": true, 00:13:33.929 "claim_type": "exclusive_write", 00:13:33.929 "zoned": false, 00:13:33.929 "supported_io_types": { 00:13:33.929 "read": true, 00:13:33.929 "write": true, 00:13:33.929 "unmap": true, 00:13:33.929 "flush": true, 00:13:33.929 "reset": true, 00:13:33.929 "nvme_admin": false, 00:13:33.929 "nvme_io": false, 00:13:33.929 "nvme_io_md": false, 00:13:33.929 "write_zeroes": true, 00:13:33.929 "zcopy": true, 00:13:33.929 "get_zone_info": false, 00:13:33.929 "zone_management": false, 00:13:33.929 "zone_append": false, 00:13:33.929 "compare": false, 00:13:33.929 "compare_and_write": false, 00:13:33.929 "abort": true, 00:13:33.929 "seek_hole": false, 00:13:33.929 "seek_data": false, 00:13:33.929 "copy": true, 00:13:33.929 "nvme_iov_md": false 00:13:33.929 }, 00:13:33.929 "memory_domains": [ 00:13:33.929 { 00:13:33.929 "dma_device_id": "system", 00:13:33.929 "dma_device_type": 1 00:13:33.929 }, 00:13:33.929 { 00:13:33.929 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:33.929 "dma_device_type": 2 00:13:33.929 } 00:13:33.929 ], 00:13:33.929 "driver_specific": {} 00:13:33.929 } 00:13:33.929 ] 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:33.929 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:33.930 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:33.930 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:33.930 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:33.930 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:33.930 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:34.188 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.188 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:34.188 "name": "Existed_Raid", 00:13:34.188 "uuid": "4eb50b6c-05c6-4ede-a8a8-92ee536525a6", 00:13:34.188 "strip_size_kb": 64, 00:13:34.188 "state": "online", 00:13:34.188 "raid_level": "raid5f", 00:13:34.188 "superblock": true, 00:13:34.188 "num_base_bdevs": 3, 00:13:34.188 "num_base_bdevs_discovered": 3, 00:13:34.188 "num_base_bdevs_operational": 3, 00:13:34.188 "base_bdevs_list": [ 00:13:34.188 { 00:13:34.188 "name": "BaseBdev1", 00:13:34.188 "uuid": "0a55f15a-0deb-44dc-8c17-3f0219c32465", 00:13:34.188 "is_configured": true, 00:13:34.188 "data_offset": 2048, 00:13:34.188 "data_size": 63488 00:13:34.188 }, 00:13:34.188 { 00:13:34.188 "name": "BaseBdev2", 00:13:34.188 "uuid": "4f6530a5-a639-4dcb-a98d-e9dfa95e7bb0", 00:13:34.188 "is_configured": true, 00:13:34.188 "data_offset": 2048, 00:13:34.188 "data_size": 63488 00:13:34.188 }, 00:13:34.188 { 00:13:34.188 "name": "BaseBdev3", 00:13:34.188 "uuid": "97d988b6-4370-4582-84a9-714d52885e84", 00:13:34.188 "is_configured": true, 00:13:34.188 "data_offset": 2048, 00:13:34.188 "data_size": 63488 00:13:34.188 } 00:13:34.188 ] 00:13:34.188 }' 00:13:34.188 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:34.188 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:34.447 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:13:34.447 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:34.447 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:34.447 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:34.447 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:13:34.447 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:34.447 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:34.447 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:34.447 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.447 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:34.447 [2024-12-05 19:03:51.907635] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:34.447 19:03:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.447 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:34.447 "name": "Existed_Raid", 00:13:34.447 "aliases": [ 00:13:34.447 "4eb50b6c-05c6-4ede-a8a8-92ee536525a6" 00:13:34.447 ], 00:13:34.447 "product_name": "Raid Volume", 00:13:34.447 "block_size": 512, 00:13:34.447 "num_blocks": 126976, 00:13:34.447 "uuid": "4eb50b6c-05c6-4ede-a8a8-92ee536525a6", 00:13:34.447 "assigned_rate_limits": { 00:13:34.447 "rw_ios_per_sec": 0, 00:13:34.447 "rw_mbytes_per_sec": 0, 00:13:34.447 "r_mbytes_per_sec": 0, 00:13:34.447 "w_mbytes_per_sec": 0 00:13:34.447 }, 00:13:34.447 "claimed": false, 00:13:34.447 "zoned": false, 00:13:34.447 "supported_io_types": { 00:13:34.447 "read": true, 00:13:34.447 "write": true, 00:13:34.447 "unmap": false, 00:13:34.447 "flush": false, 00:13:34.447 "reset": true, 00:13:34.447 "nvme_admin": false, 00:13:34.447 "nvme_io": false, 00:13:34.447 "nvme_io_md": false, 00:13:34.447 "write_zeroes": true, 00:13:34.447 "zcopy": false, 00:13:34.447 "get_zone_info": false, 00:13:34.447 "zone_management": false, 00:13:34.447 "zone_append": false, 00:13:34.447 "compare": false, 00:13:34.447 "compare_and_write": false, 00:13:34.447 "abort": false, 00:13:34.447 "seek_hole": false, 00:13:34.447 "seek_data": false, 00:13:34.447 "copy": false, 00:13:34.447 "nvme_iov_md": false 00:13:34.447 }, 00:13:34.447 "driver_specific": { 00:13:34.447 "raid": { 00:13:34.447 "uuid": "4eb50b6c-05c6-4ede-a8a8-92ee536525a6", 00:13:34.447 "strip_size_kb": 64, 00:13:34.447 "state": "online", 00:13:34.447 "raid_level": "raid5f", 00:13:34.447 "superblock": true, 00:13:34.447 "num_base_bdevs": 3, 00:13:34.447 "num_base_bdevs_discovered": 3, 00:13:34.447 "num_base_bdevs_operational": 3, 00:13:34.447 "base_bdevs_list": [ 00:13:34.447 { 00:13:34.447 "name": "BaseBdev1", 00:13:34.447 "uuid": "0a55f15a-0deb-44dc-8c17-3f0219c32465", 00:13:34.447 "is_configured": true, 00:13:34.447 "data_offset": 2048, 00:13:34.447 "data_size": 63488 00:13:34.447 }, 00:13:34.447 { 00:13:34.447 "name": "BaseBdev2", 00:13:34.447 "uuid": "4f6530a5-a639-4dcb-a98d-e9dfa95e7bb0", 00:13:34.447 "is_configured": true, 00:13:34.448 "data_offset": 2048, 00:13:34.448 "data_size": 63488 00:13:34.448 }, 00:13:34.448 { 00:13:34.448 "name": "BaseBdev3", 00:13:34.448 "uuid": "97d988b6-4370-4582-84a9-714d52885e84", 00:13:34.448 "is_configured": true, 00:13:34.448 "data_offset": 2048, 00:13:34.448 "data_size": 63488 00:13:34.448 } 00:13:34.448 ] 00:13:34.448 } 00:13:34.448 } 00:13:34.448 }' 00:13:34.448 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:34.448 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:13:34.448 BaseBdev2 00:13:34.448 BaseBdev3' 00:13:34.448 19:03:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:34.707 [2024-12-05 19:03:52.187024] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:34.707 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:34.707 "name": "Existed_Raid", 00:13:34.707 "uuid": "4eb50b6c-05c6-4ede-a8a8-92ee536525a6", 00:13:34.707 "strip_size_kb": 64, 00:13:34.707 "state": "online", 00:13:34.707 "raid_level": "raid5f", 00:13:34.707 "superblock": true, 00:13:34.707 "num_base_bdevs": 3, 00:13:34.707 "num_base_bdevs_discovered": 2, 00:13:34.707 "num_base_bdevs_operational": 2, 00:13:34.707 "base_bdevs_list": [ 00:13:34.707 { 00:13:34.707 "name": null, 00:13:34.707 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:34.707 "is_configured": false, 00:13:34.707 "data_offset": 0, 00:13:34.707 "data_size": 63488 00:13:34.707 }, 00:13:34.707 { 00:13:34.707 "name": "BaseBdev2", 00:13:34.707 "uuid": "4f6530a5-a639-4dcb-a98d-e9dfa95e7bb0", 00:13:34.707 "is_configured": true, 00:13:34.707 "data_offset": 2048, 00:13:34.707 "data_size": 63488 00:13:34.707 }, 00:13:34.707 { 00:13:34.707 "name": "BaseBdev3", 00:13:34.707 "uuid": "97d988b6-4370-4582-84a9-714d52885e84", 00:13:34.707 "is_configured": true, 00:13:34.707 "data_offset": 2048, 00:13:34.708 "data_size": 63488 00:13:34.708 } 00:13:34.708 ] 00:13:34.708 }' 00:13:34.708 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:34.708 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.275 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:13:35.275 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:35.275 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:35.275 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.275 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.276 [2024-12-05 19:03:52.625478] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:35.276 [2024-12-05 19:03:52.625599] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:35.276 [2024-12-05 19:03:52.636607] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.276 [2024-12-05 19:03:52.692534] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:35.276 [2024-12-05 19:03:52.692627] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.276 BaseBdev2 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.276 [ 00:13:35.276 { 00:13:35.276 "name": "BaseBdev2", 00:13:35.276 "aliases": [ 00:13:35.276 "12b4a6f1-1323-4ed2-b22c-e61085eae4fa" 00:13:35.276 ], 00:13:35.276 "product_name": "Malloc disk", 00:13:35.276 "block_size": 512, 00:13:35.276 "num_blocks": 65536, 00:13:35.276 "uuid": "12b4a6f1-1323-4ed2-b22c-e61085eae4fa", 00:13:35.276 "assigned_rate_limits": { 00:13:35.276 "rw_ios_per_sec": 0, 00:13:35.276 "rw_mbytes_per_sec": 0, 00:13:35.276 "r_mbytes_per_sec": 0, 00:13:35.276 "w_mbytes_per_sec": 0 00:13:35.276 }, 00:13:35.276 "claimed": false, 00:13:35.276 "zoned": false, 00:13:35.276 "supported_io_types": { 00:13:35.276 "read": true, 00:13:35.276 "write": true, 00:13:35.276 "unmap": true, 00:13:35.276 "flush": true, 00:13:35.276 "reset": true, 00:13:35.276 "nvme_admin": false, 00:13:35.276 "nvme_io": false, 00:13:35.276 "nvme_io_md": false, 00:13:35.276 "write_zeroes": true, 00:13:35.276 "zcopy": true, 00:13:35.276 "get_zone_info": false, 00:13:35.276 "zone_management": false, 00:13:35.276 "zone_append": false, 00:13:35.276 "compare": false, 00:13:35.276 "compare_and_write": false, 00:13:35.276 "abort": true, 00:13:35.276 "seek_hole": false, 00:13:35.276 "seek_data": false, 00:13:35.276 "copy": true, 00:13:35.276 "nvme_iov_md": false 00:13:35.276 }, 00:13:35.276 "memory_domains": [ 00:13:35.276 { 00:13:35.276 "dma_device_id": "system", 00:13:35.276 "dma_device_type": 1 00:13:35.276 }, 00:13:35.276 { 00:13:35.276 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:35.276 "dma_device_type": 2 00:13:35.276 } 00:13:35.276 ], 00:13:35.276 "driver_specific": {} 00:13:35.276 } 00:13:35.276 ] 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.276 BaseBdev3 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.276 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.559 [ 00:13:35.559 { 00:13:35.559 "name": "BaseBdev3", 00:13:35.559 "aliases": [ 00:13:35.559 "db9b2939-afa1-457a-9c00-cda66b1aa770" 00:13:35.559 ], 00:13:35.559 "product_name": "Malloc disk", 00:13:35.559 "block_size": 512, 00:13:35.559 "num_blocks": 65536, 00:13:35.559 "uuid": "db9b2939-afa1-457a-9c00-cda66b1aa770", 00:13:35.559 "assigned_rate_limits": { 00:13:35.559 "rw_ios_per_sec": 0, 00:13:35.559 "rw_mbytes_per_sec": 0, 00:13:35.559 "r_mbytes_per_sec": 0, 00:13:35.559 "w_mbytes_per_sec": 0 00:13:35.559 }, 00:13:35.559 "claimed": false, 00:13:35.559 "zoned": false, 00:13:35.559 "supported_io_types": { 00:13:35.559 "read": true, 00:13:35.559 "write": true, 00:13:35.559 "unmap": true, 00:13:35.559 "flush": true, 00:13:35.559 "reset": true, 00:13:35.559 "nvme_admin": false, 00:13:35.559 "nvme_io": false, 00:13:35.559 "nvme_io_md": false, 00:13:35.559 "write_zeroes": true, 00:13:35.559 "zcopy": true, 00:13:35.559 "get_zone_info": false, 00:13:35.559 "zone_management": false, 00:13:35.559 "zone_append": false, 00:13:35.559 "compare": false, 00:13:35.559 "compare_and_write": false, 00:13:35.559 "abort": true, 00:13:35.559 "seek_hole": false, 00:13:35.559 "seek_data": false, 00:13:35.559 "copy": true, 00:13:35.559 "nvme_iov_md": false 00:13:35.559 }, 00:13:35.559 "memory_domains": [ 00:13:35.559 { 00:13:35.559 "dma_device_id": "system", 00:13:35.559 "dma_device_type": 1 00:13:35.559 }, 00:13:35.559 { 00:13:35.559 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:35.559 "dma_device_type": 2 00:13:35.559 } 00:13:35.559 ], 00:13:35.559 "driver_specific": {} 00:13:35.559 } 00:13:35.559 ] 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.559 [2024-12-05 19:03:52.867324] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:35.559 [2024-12-05 19:03:52.867456] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:35.559 [2024-12-05 19:03:52.867497] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:35.559 [2024-12-05 19:03:52.869325] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:35.559 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:35.560 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:35.560 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:35.560 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:35.560 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.560 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:35.560 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.560 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.560 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.560 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:35.560 "name": "Existed_Raid", 00:13:35.560 "uuid": "27189dd4-9799-4f9a-b1c2-3accb38588e1", 00:13:35.560 "strip_size_kb": 64, 00:13:35.560 "state": "configuring", 00:13:35.560 "raid_level": "raid5f", 00:13:35.560 "superblock": true, 00:13:35.560 "num_base_bdevs": 3, 00:13:35.560 "num_base_bdevs_discovered": 2, 00:13:35.560 "num_base_bdevs_operational": 3, 00:13:35.560 "base_bdevs_list": [ 00:13:35.560 { 00:13:35.560 "name": "BaseBdev1", 00:13:35.560 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:35.560 "is_configured": false, 00:13:35.560 "data_offset": 0, 00:13:35.560 "data_size": 0 00:13:35.560 }, 00:13:35.560 { 00:13:35.560 "name": "BaseBdev2", 00:13:35.560 "uuid": "12b4a6f1-1323-4ed2-b22c-e61085eae4fa", 00:13:35.560 "is_configured": true, 00:13:35.560 "data_offset": 2048, 00:13:35.560 "data_size": 63488 00:13:35.560 }, 00:13:35.560 { 00:13:35.560 "name": "BaseBdev3", 00:13:35.560 "uuid": "db9b2939-afa1-457a-9c00-cda66b1aa770", 00:13:35.560 "is_configured": true, 00:13:35.560 "data_offset": 2048, 00:13:35.560 "data_size": 63488 00:13:35.560 } 00:13:35.560 ] 00:13:35.560 }' 00:13:35.560 19:03:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:35.560 19:03:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.820 [2024-12-05 19:03:53.310572] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:35.820 "name": "Existed_Raid", 00:13:35.820 "uuid": "27189dd4-9799-4f9a-b1c2-3accb38588e1", 00:13:35.820 "strip_size_kb": 64, 00:13:35.820 "state": "configuring", 00:13:35.820 "raid_level": "raid5f", 00:13:35.820 "superblock": true, 00:13:35.820 "num_base_bdevs": 3, 00:13:35.820 "num_base_bdevs_discovered": 1, 00:13:35.820 "num_base_bdevs_operational": 3, 00:13:35.820 "base_bdevs_list": [ 00:13:35.820 { 00:13:35.820 "name": "BaseBdev1", 00:13:35.820 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:35.820 "is_configured": false, 00:13:35.820 "data_offset": 0, 00:13:35.820 "data_size": 0 00:13:35.820 }, 00:13:35.820 { 00:13:35.820 "name": null, 00:13:35.820 "uuid": "12b4a6f1-1323-4ed2-b22c-e61085eae4fa", 00:13:35.820 "is_configured": false, 00:13:35.820 "data_offset": 0, 00:13:35.820 "data_size": 63488 00:13:35.820 }, 00:13:35.820 { 00:13:35.820 "name": "BaseBdev3", 00:13:35.820 "uuid": "db9b2939-afa1-457a-9c00-cda66b1aa770", 00:13:35.820 "is_configured": true, 00:13:35.820 "data_offset": 2048, 00:13:35.820 "data_size": 63488 00:13:35.820 } 00:13:35.820 ] 00:13:35.820 }' 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:35.820 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:36.390 [2024-12-05 19:03:53.804790] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:36.390 BaseBdev1 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:36.390 [ 00:13:36.390 { 00:13:36.390 "name": "BaseBdev1", 00:13:36.390 "aliases": [ 00:13:36.390 "835c6b71-b3c5-4db2-a017-dfd7040017b3" 00:13:36.390 ], 00:13:36.390 "product_name": "Malloc disk", 00:13:36.390 "block_size": 512, 00:13:36.390 "num_blocks": 65536, 00:13:36.390 "uuid": "835c6b71-b3c5-4db2-a017-dfd7040017b3", 00:13:36.390 "assigned_rate_limits": { 00:13:36.390 "rw_ios_per_sec": 0, 00:13:36.390 "rw_mbytes_per_sec": 0, 00:13:36.390 "r_mbytes_per_sec": 0, 00:13:36.390 "w_mbytes_per_sec": 0 00:13:36.390 }, 00:13:36.390 "claimed": true, 00:13:36.390 "claim_type": "exclusive_write", 00:13:36.390 "zoned": false, 00:13:36.390 "supported_io_types": { 00:13:36.390 "read": true, 00:13:36.390 "write": true, 00:13:36.390 "unmap": true, 00:13:36.390 "flush": true, 00:13:36.390 "reset": true, 00:13:36.390 "nvme_admin": false, 00:13:36.390 "nvme_io": false, 00:13:36.390 "nvme_io_md": false, 00:13:36.390 "write_zeroes": true, 00:13:36.390 "zcopy": true, 00:13:36.390 "get_zone_info": false, 00:13:36.390 "zone_management": false, 00:13:36.390 "zone_append": false, 00:13:36.390 "compare": false, 00:13:36.390 "compare_and_write": false, 00:13:36.390 "abort": true, 00:13:36.390 "seek_hole": false, 00:13:36.390 "seek_data": false, 00:13:36.390 "copy": true, 00:13:36.390 "nvme_iov_md": false 00:13:36.390 }, 00:13:36.390 "memory_domains": [ 00:13:36.390 { 00:13:36.390 "dma_device_id": "system", 00:13:36.390 "dma_device_type": 1 00:13:36.390 }, 00:13:36.390 { 00:13:36.390 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:36.390 "dma_device_type": 2 00:13:36.390 } 00:13:36.390 ], 00:13:36.390 "driver_specific": {} 00:13:36.390 } 00:13:36.390 ] 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:36.390 "name": "Existed_Raid", 00:13:36.390 "uuid": "27189dd4-9799-4f9a-b1c2-3accb38588e1", 00:13:36.390 "strip_size_kb": 64, 00:13:36.390 "state": "configuring", 00:13:36.390 "raid_level": "raid5f", 00:13:36.390 "superblock": true, 00:13:36.390 "num_base_bdevs": 3, 00:13:36.390 "num_base_bdevs_discovered": 2, 00:13:36.390 "num_base_bdevs_operational": 3, 00:13:36.390 "base_bdevs_list": [ 00:13:36.390 { 00:13:36.390 "name": "BaseBdev1", 00:13:36.390 "uuid": "835c6b71-b3c5-4db2-a017-dfd7040017b3", 00:13:36.390 "is_configured": true, 00:13:36.390 "data_offset": 2048, 00:13:36.390 "data_size": 63488 00:13:36.390 }, 00:13:36.390 { 00:13:36.390 "name": null, 00:13:36.390 "uuid": "12b4a6f1-1323-4ed2-b22c-e61085eae4fa", 00:13:36.390 "is_configured": false, 00:13:36.390 "data_offset": 0, 00:13:36.390 "data_size": 63488 00:13:36.390 }, 00:13:36.390 { 00:13:36.390 "name": "BaseBdev3", 00:13:36.390 "uuid": "db9b2939-afa1-457a-9c00-cda66b1aa770", 00:13:36.390 "is_configured": true, 00:13:36.390 "data_offset": 2048, 00:13:36.390 "data_size": 63488 00:13:36.390 } 00:13:36.390 ] 00:13:36.390 }' 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:36.390 19:03:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:36.960 [2024-12-05 19:03:54.335945] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:36.960 "name": "Existed_Raid", 00:13:36.960 "uuid": "27189dd4-9799-4f9a-b1c2-3accb38588e1", 00:13:36.960 "strip_size_kb": 64, 00:13:36.960 "state": "configuring", 00:13:36.960 "raid_level": "raid5f", 00:13:36.960 "superblock": true, 00:13:36.960 "num_base_bdevs": 3, 00:13:36.960 "num_base_bdevs_discovered": 1, 00:13:36.960 "num_base_bdevs_operational": 3, 00:13:36.960 "base_bdevs_list": [ 00:13:36.960 { 00:13:36.960 "name": "BaseBdev1", 00:13:36.960 "uuid": "835c6b71-b3c5-4db2-a017-dfd7040017b3", 00:13:36.960 "is_configured": true, 00:13:36.960 "data_offset": 2048, 00:13:36.960 "data_size": 63488 00:13:36.960 }, 00:13:36.960 { 00:13:36.960 "name": null, 00:13:36.960 "uuid": "12b4a6f1-1323-4ed2-b22c-e61085eae4fa", 00:13:36.960 "is_configured": false, 00:13:36.960 "data_offset": 0, 00:13:36.960 "data_size": 63488 00:13:36.960 }, 00:13:36.960 { 00:13:36.960 "name": null, 00:13:36.960 "uuid": "db9b2939-afa1-457a-9c00-cda66b1aa770", 00:13:36.960 "is_configured": false, 00:13:36.960 "data_offset": 0, 00:13:36.960 "data_size": 63488 00:13:36.960 } 00:13:36.960 ] 00:13:36.960 }' 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:36.960 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:37.530 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:37.530 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.530 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.530 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:37.530 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.530 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:37.531 [2024-12-05 19:03:54.835120] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:37.531 "name": "Existed_Raid", 00:13:37.531 "uuid": "27189dd4-9799-4f9a-b1c2-3accb38588e1", 00:13:37.531 "strip_size_kb": 64, 00:13:37.531 "state": "configuring", 00:13:37.531 "raid_level": "raid5f", 00:13:37.531 "superblock": true, 00:13:37.531 "num_base_bdevs": 3, 00:13:37.531 "num_base_bdevs_discovered": 2, 00:13:37.531 "num_base_bdevs_operational": 3, 00:13:37.531 "base_bdevs_list": [ 00:13:37.531 { 00:13:37.531 "name": "BaseBdev1", 00:13:37.531 "uuid": "835c6b71-b3c5-4db2-a017-dfd7040017b3", 00:13:37.531 "is_configured": true, 00:13:37.531 "data_offset": 2048, 00:13:37.531 "data_size": 63488 00:13:37.531 }, 00:13:37.531 { 00:13:37.531 "name": null, 00:13:37.531 "uuid": "12b4a6f1-1323-4ed2-b22c-e61085eae4fa", 00:13:37.531 "is_configured": false, 00:13:37.531 "data_offset": 0, 00:13:37.531 "data_size": 63488 00:13:37.531 }, 00:13:37.531 { 00:13:37.531 "name": "BaseBdev3", 00:13:37.531 "uuid": "db9b2939-afa1-457a-9c00-cda66b1aa770", 00:13:37.531 "is_configured": true, 00:13:37.531 "data_offset": 2048, 00:13:37.531 "data_size": 63488 00:13:37.531 } 00:13:37.531 ] 00:13:37.531 }' 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:37.531 19:03:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:37.795 [2024-12-05 19:03:55.294347] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:37.795 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.056 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:38.056 "name": "Existed_Raid", 00:13:38.056 "uuid": "27189dd4-9799-4f9a-b1c2-3accb38588e1", 00:13:38.056 "strip_size_kb": 64, 00:13:38.056 "state": "configuring", 00:13:38.056 "raid_level": "raid5f", 00:13:38.056 "superblock": true, 00:13:38.056 "num_base_bdevs": 3, 00:13:38.056 "num_base_bdevs_discovered": 1, 00:13:38.056 "num_base_bdevs_operational": 3, 00:13:38.056 "base_bdevs_list": [ 00:13:38.056 { 00:13:38.056 "name": null, 00:13:38.056 "uuid": "835c6b71-b3c5-4db2-a017-dfd7040017b3", 00:13:38.056 "is_configured": false, 00:13:38.056 "data_offset": 0, 00:13:38.056 "data_size": 63488 00:13:38.056 }, 00:13:38.056 { 00:13:38.056 "name": null, 00:13:38.056 "uuid": "12b4a6f1-1323-4ed2-b22c-e61085eae4fa", 00:13:38.056 "is_configured": false, 00:13:38.056 "data_offset": 0, 00:13:38.056 "data_size": 63488 00:13:38.056 }, 00:13:38.056 { 00:13:38.056 "name": "BaseBdev3", 00:13:38.056 "uuid": "db9b2939-afa1-457a-9c00-cda66b1aa770", 00:13:38.056 "is_configured": true, 00:13:38.056 "data_offset": 2048, 00:13:38.056 "data_size": 63488 00:13:38.056 } 00:13:38.056 ] 00:13:38.056 }' 00:13:38.056 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:38.056 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:38.315 [2024-12-05 19:03:55.775890] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:38.315 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:38.316 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:38.316 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:38.316 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:38.316 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.316 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:38.316 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.316 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:38.316 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.316 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:38.316 "name": "Existed_Raid", 00:13:38.316 "uuid": "27189dd4-9799-4f9a-b1c2-3accb38588e1", 00:13:38.316 "strip_size_kb": 64, 00:13:38.316 "state": "configuring", 00:13:38.316 "raid_level": "raid5f", 00:13:38.316 "superblock": true, 00:13:38.316 "num_base_bdevs": 3, 00:13:38.316 "num_base_bdevs_discovered": 2, 00:13:38.316 "num_base_bdevs_operational": 3, 00:13:38.316 "base_bdevs_list": [ 00:13:38.316 { 00:13:38.316 "name": null, 00:13:38.316 "uuid": "835c6b71-b3c5-4db2-a017-dfd7040017b3", 00:13:38.316 "is_configured": false, 00:13:38.316 "data_offset": 0, 00:13:38.316 "data_size": 63488 00:13:38.316 }, 00:13:38.316 { 00:13:38.316 "name": "BaseBdev2", 00:13:38.316 "uuid": "12b4a6f1-1323-4ed2-b22c-e61085eae4fa", 00:13:38.316 "is_configured": true, 00:13:38.316 "data_offset": 2048, 00:13:38.316 "data_size": 63488 00:13:38.316 }, 00:13:38.316 { 00:13:38.316 "name": "BaseBdev3", 00:13:38.316 "uuid": "db9b2939-afa1-457a-9c00-cda66b1aa770", 00:13:38.316 "is_configured": true, 00:13:38.316 "data_offset": 2048, 00:13:38.316 "data_size": 63488 00:13:38.316 } 00:13:38.316 ] 00:13:38.316 }' 00:13:38.316 19:03:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:38.316 19:03:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 835c6b71-b3c5-4db2-a017-dfd7040017b3 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:38.885 [2024-12-05 19:03:56.321312] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:13:38.885 [2024-12-05 19:03:56.321558] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:13:38.885 [2024-12-05 19:03:56.321614] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:38.885 [2024-12-05 19:03:56.321898] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:13:38.885 NewBaseBdev 00:13:38.885 [2024-12-05 19:03:56.322316] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:13:38.885 [2024-12-05 19:03:56.322374] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:13:38.885 [2024-12-05 19:03:56.322514] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.885 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:38.885 [ 00:13:38.885 { 00:13:38.885 "name": "NewBaseBdev", 00:13:38.885 "aliases": [ 00:13:38.885 "835c6b71-b3c5-4db2-a017-dfd7040017b3" 00:13:38.885 ], 00:13:38.885 "product_name": "Malloc disk", 00:13:38.885 "block_size": 512, 00:13:38.885 "num_blocks": 65536, 00:13:38.885 "uuid": "835c6b71-b3c5-4db2-a017-dfd7040017b3", 00:13:38.885 "assigned_rate_limits": { 00:13:38.885 "rw_ios_per_sec": 0, 00:13:38.885 "rw_mbytes_per_sec": 0, 00:13:38.885 "r_mbytes_per_sec": 0, 00:13:38.885 "w_mbytes_per_sec": 0 00:13:38.885 }, 00:13:38.885 "claimed": true, 00:13:38.885 "claim_type": "exclusive_write", 00:13:38.885 "zoned": false, 00:13:38.885 "supported_io_types": { 00:13:38.885 "read": true, 00:13:38.885 "write": true, 00:13:38.885 "unmap": true, 00:13:38.885 "flush": true, 00:13:38.885 "reset": true, 00:13:38.885 "nvme_admin": false, 00:13:38.885 "nvme_io": false, 00:13:38.885 "nvme_io_md": false, 00:13:38.885 "write_zeroes": true, 00:13:38.885 "zcopy": true, 00:13:38.885 "get_zone_info": false, 00:13:38.885 "zone_management": false, 00:13:38.885 "zone_append": false, 00:13:38.885 "compare": false, 00:13:38.885 "compare_and_write": false, 00:13:38.885 "abort": true, 00:13:38.885 "seek_hole": false, 00:13:38.885 "seek_data": false, 00:13:38.885 "copy": true, 00:13:38.886 "nvme_iov_md": false 00:13:38.886 }, 00:13:38.886 "memory_domains": [ 00:13:38.886 { 00:13:38.886 "dma_device_id": "system", 00:13:38.886 "dma_device_type": 1 00:13:38.886 }, 00:13:38.886 { 00:13:38.886 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:38.886 "dma_device_type": 2 00:13:38.886 } 00:13:38.886 ], 00:13:38.886 "driver_specific": {} 00:13:38.886 } 00:13:38.886 ] 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:38.886 "name": "Existed_Raid", 00:13:38.886 "uuid": "27189dd4-9799-4f9a-b1c2-3accb38588e1", 00:13:38.886 "strip_size_kb": 64, 00:13:38.886 "state": "online", 00:13:38.886 "raid_level": "raid5f", 00:13:38.886 "superblock": true, 00:13:38.886 "num_base_bdevs": 3, 00:13:38.886 "num_base_bdevs_discovered": 3, 00:13:38.886 "num_base_bdevs_operational": 3, 00:13:38.886 "base_bdevs_list": [ 00:13:38.886 { 00:13:38.886 "name": "NewBaseBdev", 00:13:38.886 "uuid": "835c6b71-b3c5-4db2-a017-dfd7040017b3", 00:13:38.886 "is_configured": true, 00:13:38.886 "data_offset": 2048, 00:13:38.886 "data_size": 63488 00:13:38.886 }, 00:13:38.886 { 00:13:38.886 "name": "BaseBdev2", 00:13:38.886 "uuid": "12b4a6f1-1323-4ed2-b22c-e61085eae4fa", 00:13:38.886 "is_configured": true, 00:13:38.886 "data_offset": 2048, 00:13:38.886 "data_size": 63488 00:13:38.886 }, 00:13:38.886 { 00:13:38.886 "name": "BaseBdev3", 00:13:38.886 "uuid": "db9b2939-afa1-457a-9c00-cda66b1aa770", 00:13:38.886 "is_configured": true, 00:13:38.886 "data_offset": 2048, 00:13:38.886 "data_size": 63488 00:13:38.886 } 00:13:38.886 ] 00:13:38.886 }' 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:38.886 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.457 [2024-12-05 19:03:56.784710] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:39.457 "name": "Existed_Raid", 00:13:39.457 "aliases": [ 00:13:39.457 "27189dd4-9799-4f9a-b1c2-3accb38588e1" 00:13:39.457 ], 00:13:39.457 "product_name": "Raid Volume", 00:13:39.457 "block_size": 512, 00:13:39.457 "num_blocks": 126976, 00:13:39.457 "uuid": "27189dd4-9799-4f9a-b1c2-3accb38588e1", 00:13:39.457 "assigned_rate_limits": { 00:13:39.457 "rw_ios_per_sec": 0, 00:13:39.457 "rw_mbytes_per_sec": 0, 00:13:39.457 "r_mbytes_per_sec": 0, 00:13:39.457 "w_mbytes_per_sec": 0 00:13:39.457 }, 00:13:39.457 "claimed": false, 00:13:39.457 "zoned": false, 00:13:39.457 "supported_io_types": { 00:13:39.457 "read": true, 00:13:39.457 "write": true, 00:13:39.457 "unmap": false, 00:13:39.457 "flush": false, 00:13:39.457 "reset": true, 00:13:39.457 "nvme_admin": false, 00:13:39.457 "nvme_io": false, 00:13:39.457 "nvme_io_md": false, 00:13:39.457 "write_zeroes": true, 00:13:39.457 "zcopy": false, 00:13:39.457 "get_zone_info": false, 00:13:39.457 "zone_management": false, 00:13:39.457 "zone_append": false, 00:13:39.457 "compare": false, 00:13:39.457 "compare_and_write": false, 00:13:39.457 "abort": false, 00:13:39.457 "seek_hole": false, 00:13:39.457 "seek_data": false, 00:13:39.457 "copy": false, 00:13:39.457 "nvme_iov_md": false 00:13:39.457 }, 00:13:39.457 "driver_specific": { 00:13:39.457 "raid": { 00:13:39.457 "uuid": "27189dd4-9799-4f9a-b1c2-3accb38588e1", 00:13:39.457 "strip_size_kb": 64, 00:13:39.457 "state": "online", 00:13:39.457 "raid_level": "raid5f", 00:13:39.457 "superblock": true, 00:13:39.457 "num_base_bdevs": 3, 00:13:39.457 "num_base_bdevs_discovered": 3, 00:13:39.457 "num_base_bdevs_operational": 3, 00:13:39.457 "base_bdevs_list": [ 00:13:39.457 { 00:13:39.457 "name": "NewBaseBdev", 00:13:39.457 "uuid": "835c6b71-b3c5-4db2-a017-dfd7040017b3", 00:13:39.457 "is_configured": true, 00:13:39.457 "data_offset": 2048, 00:13:39.457 "data_size": 63488 00:13:39.457 }, 00:13:39.457 { 00:13:39.457 "name": "BaseBdev2", 00:13:39.457 "uuid": "12b4a6f1-1323-4ed2-b22c-e61085eae4fa", 00:13:39.457 "is_configured": true, 00:13:39.457 "data_offset": 2048, 00:13:39.457 "data_size": 63488 00:13:39.457 }, 00:13:39.457 { 00:13:39.457 "name": "BaseBdev3", 00:13:39.457 "uuid": "db9b2939-afa1-457a-9c00-cda66b1aa770", 00:13:39.457 "is_configured": true, 00:13:39.457 "data_offset": 2048, 00:13:39.457 "data_size": 63488 00:13:39.457 } 00:13:39.457 ] 00:13:39.457 } 00:13:39.457 } 00:13:39.457 }' 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:13:39.457 BaseBdev2 00:13:39.457 BaseBdev3' 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:39.457 19:03:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.718 19:03:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:39.718 19:03:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:39.718 19:03:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:39.718 19:03:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:39.718 19:03:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:39.718 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.718 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.718 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.718 19:03:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.719 [2024-12-05 19:03:57.048074] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:39.719 [2024-12-05 19:03:57.048097] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:39.719 [2024-12-05 19:03:57.048156] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:39.719 [2024-12-05 19:03:57.048376] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:39.719 [2024-12-05 19:03:57.048388] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 88963 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 88963 ']' 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 88963 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 88963 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 88963' 00:13:39.719 killing process with pid 88963 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 88963 00:13:39.719 [2024-12-05 19:03:57.097342] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:39.719 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 88963 00:13:39.719 [2024-12-05 19:03:57.127751] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:39.979 ************************************ 00:13:39.979 END TEST raid5f_state_function_test_sb 00:13:39.979 ************************************ 00:13:39.979 19:03:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:13:39.979 00:13:39.979 real 0m8.748s 00:13:39.979 user 0m14.865s 00:13:39.979 sys 0m1.900s 00:13:39.979 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:39.979 19:03:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:39.979 19:03:57 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 3 00:13:39.979 19:03:57 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:13:39.979 19:03:57 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:39.979 19:03:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:39.979 ************************************ 00:13:39.979 START TEST raid5f_superblock_test 00:13:39.979 ************************************ 00:13:39.979 19:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid5f 3 00:13:39.979 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:13:39.979 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:13:39.979 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:13:39.979 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:13:39.979 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:13:39.979 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=89513 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 89513 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 89513 ']' 00:13:39.980 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:39.980 19:03:57 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.980 [2024-12-05 19:03:57.518520] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:13:39.980 [2024-12-05 19:03:57.518765] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89513 ] 00:13:40.239 [2024-12-05 19:03:57.670751] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:40.239 [2024-12-05 19:03:57.695876] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:40.239 [2024-12-05 19:03:57.739351] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:40.239 [2024-12-05 19:03:57.739475] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:40.810 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:40.810 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:13:40.810 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:13:40.810 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:40.810 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:13:40.810 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:13:40.810 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:13:40.810 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:40.810 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:40.810 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:40.810 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.811 malloc1 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:40.811 [2024-12-05 19:03:58.351594] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:40.811 [2024-12-05 19:03:58.351758] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:40.811 [2024-12-05 19:03:58.351784] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:40.811 [2024-12-05 19:03:58.351797] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:40.811 [2024-12-05 19:03:58.354024] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:40.811 [2024-12-05 19:03:58.354065] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:40.811 pt1 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:40.811 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.071 malloc2 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.071 [2024-12-05 19:03:58.380146] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:41.071 [2024-12-05 19:03:58.380201] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:41.071 [2024-12-05 19:03:58.380217] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:41.071 [2024-12-05 19:03:58.380227] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:41.071 [2024-12-05 19:03:58.382270] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:41.071 [2024-12-05 19:03:58.382307] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:41.071 pt2 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.071 malloc3 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.071 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.071 [2024-12-05 19:03:58.408725] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:41.071 [2024-12-05 19:03:58.408846] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:41.071 [2024-12-05 19:03:58.408880] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:41.071 [2024-12-05 19:03:58.408908] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:41.071 [2024-12-05 19:03:58.410994] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:41.071 [2024-12-05 19:03:58.411067] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:41.071 pt3 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.072 [2024-12-05 19:03:58.420758] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:41.072 [2024-12-05 19:03:58.422645] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:41.072 [2024-12-05 19:03:58.422746] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:41.072 [2024-12-05 19:03:58.422937] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:41.072 [2024-12-05 19:03:58.422997] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:41.072 [2024-12-05 19:03:58.423259] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:13:41.072 [2024-12-05 19:03:58.423727] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:41.072 [2024-12-05 19:03:58.423779] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:41.072 [2024-12-05 19:03:58.423940] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:41.072 "name": "raid_bdev1", 00:13:41.072 "uuid": "364d237a-2043-4b9d-a0c0-2eca9c9bb6c8", 00:13:41.072 "strip_size_kb": 64, 00:13:41.072 "state": "online", 00:13:41.072 "raid_level": "raid5f", 00:13:41.072 "superblock": true, 00:13:41.072 "num_base_bdevs": 3, 00:13:41.072 "num_base_bdevs_discovered": 3, 00:13:41.072 "num_base_bdevs_operational": 3, 00:13:41.072 "base_bdevs_list": [ 00:13:41.072 { 00:13:41.072 "name": "pt1", 00:13:41.072 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:41.072 "is_configured": true, 00:13:41.072 "data_offset": 2048, 00:13:41.072 "data_size": 63488 00:13:41.072 }, 00:13:41.072 { 00:13:41.072 "name": "pt2", 00:13:41.072 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:41.072 "is_configured": true, 00:13:41.072 "data_offset": 2048, 00:13:41.072 "data_size": 63488 00:13:41.072 }, 00:13:41.072 { 00:13:41.072 "name": "pt3", 00:13:41.072 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:41.072 "is_configured": true, 00:13:41.072 "data_offset": 2048, 00:13:41.072 "data_size": 63488 00:13:41.072 } 00:13:41.072 ] 00:13:41.072 }' 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:41.072 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.332 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:13:41.332 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:13:41.332 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:41.332 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:41.332 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:41.332 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:41.332 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:41.332 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:41.332 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.332 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.332 [2024-12-05 19:03:58.872698] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:41.592 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.592 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:41.592 "name": "raid_bdev1", 00:13:41.592 "aliases": [ 00:13:41.592 "364d237a-2043-4b9d-a0c0-2eca9c9bb6c8" 00:13:41.592 ], 00:13:41.592 "product_name": "Raid Volume", 00:13:41.592 "block_size": 512, 00:13:41.592 "num_blocks": 126976, 00:13:41.592 "uuid": "364d237a-2043-4b9d-a0c0-2eca9c9bb6c8", 00:13:41.592 "assigned_rate_limits": { 00:13:41.592 "rw_ios_per_sec": 0, 00:13:41.592 "rw_mbytes_per_sec": 0, 00:13:41.592 "r_mbytes_per_sec": 0, 00:13:41.592 "w_mbytes_per_sec": 0 00:13:41.592 }, 00:13:41.592 "claimed": false, 00:13:41.592 "zoned": false, 00:13:41.592 "supported_io_types": { 00:13:41.592 "read": true, 00:13:41.592 "write": true, 00:13:41.592 "unmap": false, 00:13:41.592 "flush": false, 00:13:41.592 "reset": true, 00:13:41.592 "nvme_admin": false, 00:13:41.592 "nvme_io": false, 00:13:41.592 "nvme_io_md": false, 00:13:41.592 "write_zeroes": true, 00:13:41.592 "zcopy": false, 00:13:41.592 "get_zone_info": false, 00:13:41.592 "zone_management": false, 00:13:41.592 "zone_append": false, 00:13:41.592 "compare": false, 00:13:41.592 "compare_and_write": false, 00:13:41.592 "abort": false, 00:13:41.592 "seek_hole": false, 00:13:41.592 "seek_data": false, 00:13:41.592 "copy": false, 00:13:41.592 "nvme_iov_md": false 00:13:41.592 }, 00:13:41.592 "driver_specific": { 00:13:41.592 "raid": { 00:13:41.592 "uuid": "364d237a-2043-4b9d-a0c0-2eca9c9bb6c8", 00:13:41.592 "strip_size_kb": 64, 00:13:41.592 "state": "online", 00:13:41.592 "raid_level": "raid5f", 00:13:41.592 "superblock": true, 00:13:41.592 "num_base_bdevs": 3, 00:13:41.592 "num_base_bdevs_discovered": 3, 00:13:41.592 "num_base_bdevs_operational": 3, 00:13:41.592 "base_bdevs_list": [ 00:13:41.592 { 00:13:41.592 "name": "pt1", 00:13:41.592 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:41.592 "is_configured": true, 00:13:41.593 "data_offset": 2048, 00:13:41.593 "data_size": 63488 00:13:41.593 }, 00:13:41.593 { 00:13:41.593 "name": "pt2", 00:13:41.593 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:41.593 "is_configured": true, 00:13:41.593 "data_offset": 2048, 00:13:41.593 "data_size": 63488 00:13:41.593 }, 00:13:41.593 { 00:13:41.593 "name": "pt3", 00:13:41.593 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:41.593 "is_configured": true, 00:13:41.593 "data_offset": 2048, 00:13:41.593 "data_size": 63488 00:13:41.593 } 00:13:41.593 ] 00:13:41.593 } 00:13:41.593 } 00:13:41.593 }' 00:13:41.593 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:41.593 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:13:41.593 pt2 00:13:41.593 pt3' 00:13:41.593 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:41.593 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:41.593 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:41.593 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:13:41.593 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.593 19:03:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.593 19:03:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.593 [2024-12-05 19:03:59.128241] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=364d237a-2043-4b9d-a0c0-2eca9c9bb6c8 00:13:41.593 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 364d237a-2043-4b9d-a0c0-2eca9c9bb6c8 ']' 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.854 [2024-12-05 19:03:59.156059] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:41.854 [2024-12-05 19:03:59.156082] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:41.854 [2024-12-05 19:03:59.156155] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:41.854 [2024-12-05 19:03:59.156211] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:41.854 [2024-12-05 19:03:59.156229] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.854 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.854 [2024-12-05 19:03:59.303886] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:13:41.854 [2024-12-05 19:03:59.305655] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:13:41.854 [2024-12-05 19:03:59.305720] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:13:41.854 [2024-12-05 19:03:59.305762] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:13:41.854 [2024-12-05 19:03:59.305795] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:13:41.854 [2024-12-05 19:03:59.305814] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:13:41.855 [2024-12-05 19:03:59.305826] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:41.855 [2024-12-05 19:03:59.305837] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:13:41.855 request: 00:13:41.855 { 00:13:41.855 "name": "raid_bdev1", 00:13:41.855 "raid_level": "raid5f", 00:13:41.855 "base_bdevs": [ 00:13:41.855 "malloc1", 00:13:41.855 "malloc2", 00:13:41.855 "malloc3" 00:13:41.855 ], 00:13:41.855 "strip_size_kb": 64, 00:13:41.855 "superblock": false, 00:13:41.855 "method": "bdev_raid_create", 00:13:41.855 "req_id": 1 00:13:41.855 } 00:13:41.855 Got JSON-RPC error response 00:13:41.855 response: 00:13:41.855 { 00:13:41.855 "code": -17, 00:13:41.855 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:13:41.855 } 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.855 [2024-12-05 19:03:59.371761] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:41.855 [2024-12-05 19:03:59.371853] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:41.855 [2024-12-05 19:03:59.371882] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:41.855 [2024-12-05 19:03:59.371909] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:41.855 [2024-12-05 19:03:59.373937] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:41.855 [2024-12-05 19:03:59.374008] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:41.855 [2024-12-05 19:03:59.374079] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:13:41.855 [2024-12-05 19:03:59.374160] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:41.855 pt1 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.855 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.115 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:42.115 "name": "raid_bdev1", 00:13:42.115 "uuid": "364d237a-2043-4b9d-a0c0-2eca9c9bb6c8", 00:13:42.115 "strip_size_kb": 64, 00:13:42.115 "state": "configuring", 00:13:42.115 "raid_level": "raid5f", 00:13:42.115 "superblock": true, 00:13:42.115 "num_base_bdevs": 3, 00:13:42.115 "num_base_bdevs_discovered": 1, 00:13:42.115 "num_base_bdevs_operational": 3, 00:13:42.115 "base_bdevs_list": [ 00:13:42.115 { 00:13:42.115 "name": "pt1", 00:13:42.115 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:42.115 "is_configured": true, 00:13:42.115 "data_offset": 2048, 00:13:42.115 "data_size": 63488 00:13:42.115 }, 00:13:42.115 { 00:13:42.115 "name": null, 00:13:42.115 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:42.115 "is_configured": false, 00:13:42.115 "data_offset": 2048, 00:13:42.115 "data_size": 63488 00:13:42.115 }, 00:13:42.115 { 00:13:42.115 "name": null, 00:13:42.115 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:42.115 "is_configured": false, 00:13:42.115 "data_offset": 2048, 00:13:42.115 "data_size": 63488 00:13:42.115 } 00:13:42.115 ] 00:13:42.115 }' 00:13:42.115 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:42.115 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:42.376 [2024-12-05 19:03:59.838940] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:42.376 [2024-12-05 19:03:59.838993] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:42.376 [2024-12-05 19:03:59.839007] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:13:42.376 [2024-12-05 19:03:59.839018] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:42.376 [2024-12-05 19:03:59.839318] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:42.376 [2024-12-05 19:03:59.839335] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:42.376 [2024-12-05 19:03:59.839379] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:42.376 [2024-12-05 19:03:59.839398] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:42.376 pt2 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:42.376 [2024-12-05 19:03:59.850936] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:42.376 "name": "raid_bdev1", 00:13:42.376 "uuid": "364d237a-2043-4b9d-a0c0-2eca9c9bb6c8", 00:13:42.376 "strip_size_kb": 64, 00:13:42.376 "state": "configuring", 00:13:42.376 "raid_level": "raid5f", 00:13:42.376 "superblock": true, 00:13:42.376 "num_base_bdevs": 3, 00:13:42.376 "num_base_bdevs_discovered": 1, 00:13:42.376 "num_base_bdevs_operational": 3, 00:13:42.376 "base_bdevs_list": [ 00:13:42.376 { 00:13:42.376 "name": "pt1", 00:13:42.376 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:42.376 "is_configured": true, 00:13:42.376 "data_offset": 2048, 00:13:42.376 "data_size": 63488 00:13:42.376 }, 00:13:42.376 { 00:13:42.376 "name": null, 00:13:42.376 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:42.376 "is_configured": false, 00:13:42.376 "data_offset": 0, 00:13:42.376 "data_size": 63488 00:13:42.376 }, 00:13:42.376 { 00:13:42.376 "name": null, 00:13:42.376 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:42.376 "is_configured": false, 00:13:42.376 "data_offset": 2048, 00:13:42.376 "data_size": 63488 00:13:42.376 } 00:13:42.376 ] 00:13:42.376 }' 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:42.376 19:03:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:42.947 [2024-12-05 19:04:00.290178] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:42.947 [2024-12-05 19:04:00.290227] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:42.947 [2024-12-05 19:04:00.290242] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:13:42.947 [2024-12-05 19:04:00.290250] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:42.947 [2024-12-05 19:04:00.290534] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:42.947 [2024-12-05 19:04:00.290549] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:42.947 [2024-12-05 19:04:00.290597] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:42.947 [2024-12-05 19:04:00.290612] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:42.947 pt2 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:42.947 [2024-12-05 19:04:00.302166] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:42.947 [2024-12-05 19:04:00.302253] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:42.947 [2024-12-05 19:04:00.302272] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:13:42.947 [2024-12-05 19:04:00.302280] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:42.947 [2024-12-05 19:04:00.302605] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:42.947 [2024-12-05 19:04:00.302622] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:42.947 [2024-12-05 19:04:00.302670] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:42.947 [2024-12-05 19:04:00.302685] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:42.947 [2024-12-05 19:04:00.302805] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:13:42.947 [2024-12-05 19:04:00.302823] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:42.947 [2024-12-05 19:04:00.303029] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:13:42.947 [2024-12-05 19:04:00.303392] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:13:42.947 [2024-12-05 19:04:00.303411] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:13:42.947 [2024-12-05 19:04:00.303501] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:42.947 pt3 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:42.947 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:42.947 "name": "raid_bdev1", 00:13:42.947 "uuid": "364d237a-2043-4b9d-a0c0-2eca9c9bb6c8", 00:13:42.947 "strip_size_kb": 64, 00:13:42.947 "state": "online", 00:13:42.947 "raid_level": "raid5f", 00:13:42.947 "superblock": true, 00:13:42.947 "num_base_bdevs": 3, 00:13:42.947 "num_base_bdevs_discovered": 3, 00:13:42.947 "num_base_bdevs_operational": 3, 00:13:42.947 "base_bdevs_list": [ 00:13:42.947 { 00:13:42.947 "name": "pt1", 00:13:42.948 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:42.948 "is_configured": true, 00:13:42.948 "data_offset": 2048, 00:13:42.948 "data_size": 63488 00:13:42.948 }, 00:13:42.948 { 00:13:42.948 "name": "pt2", 00:13:42.948 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:42.948 "is_configured": true, 00:13:42.948 "data_offset": 2048, 00:13:42.948 "data_size": 63488 00:13:42.948 }, 00:13:42.948 { 00:13:42.948 "name": "pt3", 00:13:42.948 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:42.948 "is_configured": true, 00:13:42.948 "data_offset": 2048, 00:13:42.948 "data_size": 63488 00:13:42.948 } 00:13:42.948 ] 00:13:42.948 }' 00:13:42.948 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:42.948 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.207 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:13:43.207 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:13:43.207 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:43.207 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:43.207 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:43.207 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:43.207 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:43.207 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:43.207 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.207 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.207 [2024-12-05 19:04:00.733626] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:43.207 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.207 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:43.207 "name": "raid_bdev1", 00:13:43.207 "aliases": [ 00:13:43.207 "364d237a-2043-4b9d-a0c0-2eca9c9bb6c8" 00:13:43.207 ], 00:13:43.207 "product_name": "Raid Volume", 00:13:43.207 "block_size": 512, 00:13:43.207 "num_blocks": 126976, 00:13:43.207 "uuid": "364d237a-2043-4b9d-a0c0-2eca9c9bb6c8", 00:13:43.207 "assigned_rate_limits": { 00:13:43.207 "rw_ios_per_sec": 0, 00:13:43.207 "rw_mbytes_per_sec": 0, 00:13:43.207 "r_mbytes_per_sec": 0, 00:13:43.207 "w_mbytes_per_sec": 0 00:13:43.207 }, 00:13:43.207 "claimed": false, 00:13:43.207 "zoned": false, 00:13:43.207 "supported_io_types": { 00:13:43.207 "read": true, 00:13:43.207 "write": true, 00:13:43.207 "unmap": false, 00:13:43.207 "flush": false, 00:13:43.207 "reset": true, 00:13:43.207 "nvme_admin": false, 00:13:43.207 "nvme_io": false, 00:13:43.207 "nvme_io_md": false, 00:13:43.207 "write_zeroes": true, 00:13:43.207 "zcopy": false, 00:13:43.207 "get_zone_info": false, 00:13:43.207 "zone_management": false, 00:13:43.207 "zone_append": false, 00:13:43.207 "compare": false, 00:13:43.207 "compare_and_write": false, 00:13:43.207 "abort": false, 00:13:43.207 "seek_hole": false, 00:13:43.207 "seek_data": false, 00:13:43.207 "copy": false, 00:13:43.207 "nvme_iov_md": false 00:13:43.207 }, 00:13:43.207 "driver_specific": { 00:13:43.207 "raid": { 00:13:43.207 "uuid": "364d237a-2043-4b9d-a0c0-2eca9c9bb6c8", 00:13:43.207 "strip_size_kb": 64, 00:13:43.207 "state": "online", 00:13:43.207 "raid_level": "raid5f", 00:13:43.207 "superblock": true, 00:13:43.207 "num_base_bdevs": 3, 00:13:43.207 "num_base_bdevs_discovered": 3, 00:13:43.207 "num_base_bdevs_operational": 3, 00:13:43.207 "base_bdevs_list": [ 00:13:43.207 { 00:13:43.207 "name": "pt1", 00:13:43.207 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:43.207 "is_configured": true, 00:13:43.207 "data_offset": 2048, 00:13:43.207 "data_size": 63488 00:13:43.207 }, 00:13:43.207 { 00:13:43.207 "name": "pt2", 00:13:43.207 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:43.207 "is_configured": true, 00:13:43.207 "data_offset": 2048, 00:13:43.207 "data_size": 63488 00:13:43.207 }, 00:13:43.207 { 00:13:43.207 "name": "pt3", 00:13:43.207 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:43.207 "is_configured": true, 00:13:43.207 "data_offset": 2048, 00:13:43.207 "data_size": 63488 00:13:43.207 } 00:13:43.207 ] 00:13:43.207 } 00:13:43.207 } 00:13:43.207 }' 00:13:43.207 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:13:43.467 pt2 00:13:43.467 pt3' 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:43.467 19:04:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.467 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:43.467 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:43.467 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:43.467 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:13:43.468 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.468 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.468 [2024-12-05 19:04:01.021106] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 364d237a-2043-4b9d-a0c0-2eca9c9bb6c8 '!=' 364d237a-2043-4b9d-a0c0-2eca9c9bb6c8 ']' 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.727 [2024-12-05 19:04:01.064912] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:43.727 "name": "raid_bdev1", 00:13:43.727 "uuid": "364d237a-2043-4b9d-a0c0-2eca9c9bb6c8", 00:13:43.727 "strip_size_kb": 64, 00:13:43.727 "state": "online", 00:13:43.727 "raid_level": "raid5f", 00:13:43.727 "superblock": true, 00:13:43.727 "num_base_bdevs": 3, 00:13:43.727 "num_base_bdevs_discovered": 2, 00:13:43.727 "num_base_bdevs_operational": 2, 00:13:43.727 "base_bdevs_list": [ 00:13:43.727 { 00:13:43.727 "name": null, 00:13:43.727 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.727 "is_configured": false, 00:13:43.727 "data_offset": 0, 00:13:43.727 "data_size": 63488 00:13:43.727 }, 00:13:43.727 { 00:13:43.727 "name": "pt2", 00:13:43.727 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:43.727 "is_configured": true, 00:13:43.727 "data_offset": 2048, 00:13:43.727 "data_size": 63488 00:13:43.727 }, 00:13:43.727 { 00:13:43.727 "name": "pt3", 00:13:43.727 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:43.727 "is_configured": true, 00:13:43.727 "data_offset": 2048, 00:13:43.727 "data_size": 63488 00:13:43.727 } 00:13:43.727 ] 00:13:43.727 }' 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:43.727 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.987 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:43.987 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.987 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.987 [2024-12-05 19:04:01.520131] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:43.987 [2024-12-05 19:04:01.520206] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:43.987 [2024-12-05 19:04:01.520270] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:43.987 [2024-12-05 19:04:01.520345] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:43.987 [2024-12-05 19:04:01.520395] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:13:43.987 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:43.987 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:13:43.987 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.988 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:43.988 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.988 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:44.247 [2024-12-05 19:04:01.592020] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:44.247 [2024-12-05 19:04:01.592065] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:44.247 [2024-12-05 19:04:01.592079] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:13:44.247 [2024-12-05 19:04:01.592086] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:44.247 [2024-12-05 19:04:01.594120] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:44.247 [2024-12-05 19:04:01.594156] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:44.247 [2024-12-05 19:04:01.594214] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:44.247 [2024-12-05 19:04:01.594243] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:44.247 pt2 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.247 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:44.247 "name": "raid_bdev1", 00:13:44.247 "uuid": "364d237a-2043-4b9d-a0c0-2eca9c9bb6c8", 00:13:44.247 "strip_size_kb": 64, 00:13:44.247 "state": "configuring", 00:13:44.247 "raid_level": "raid5f", 00:13:44.247 "superblock": true, 00:13:44.247 "num_base_bdevs": 3, 00:13:44.247 "num_base_bdevs_discovered": 1, 00:13:44.247 "num_base_bdevs_operational": 2, 00:13:44.247 "base_bdevs_list": [ 00:13:44.247 { 00:13:44.247 "name": null, 00:13:44.247 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:44.247 "is_configured": false, 00:13:44.247 "data_offset": 2048, 00:13:44.247 "data_size": 63488 00:13:44.247 }, 00:13:44.247 { 00:13:44.247 "name": "pt2", 00:13:44.247 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:44.248 "is_configured": true, 00:13:44.248 "data_offset": 2048, 00:13:44.248 "data_size": 63488 00:13:44.248 }, 00:13:44.248 { 00:13:44.248 "name": null, 00:13:44.248 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:44.248 "is_configured": false, 00:13:44.248 "data_offset": 2048, 00:13:44.248 "data_size": 63488 00:13:44.248 } 00:13:44.248 ] 00:13:44.248 }' 00:13:44.248 19:04:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:44.248 19:04:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:44.507 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:13:44.507 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:13:44.507 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:13:44.507 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:44.507 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.507 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:44.507 [2024-12-05 19:04:02.063214] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:44.507 [2024-12-05 19:04:02.063325] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:44.507 [2024-12-05 19:04:02.063363] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:13:44.507 [2024-12-05 19:04:02.063393] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:44.507 [2024-12-05 19:04:02.063738] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:44.507 [2024-12-05 19:04:02.063791] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:44.507 [2024-12-05 19:04:02.063868] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:44.507 [2024-12-05 19:04:02.063911] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:44.507 [2024-12-05 19:04:02.064016] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:13:44.507 [2024-12-05 19:04:02.064052] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:44.507 [2024-12-05 19:04:02.064291] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:13:44.507 [2024-12-05 19:04:02.064775] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:13:44.507 [2024-12-05 19:04:02.064829] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:13:44.507 [2024-12-05 19:04:02.065075] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:44.768 pt3 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:44.768 "name": "raid_bdev1", 00:13:44.768 "uuid": "364d237a-2043-4b9d-a0c0-2eca9c9bb6c8", 00:13:44.768 "strip_size_kb": 64, 00:13:44.768 "state": "online", 00:13:44.768 "raid_level": "raid5f", 00:13:44.768 "superblock": true, 00:13:44.768 "num_base_bdevs": 3, 00:13:44.768 "num_base_bdevs_discovered": 2, 00:13:44.768 "num_base_bdevs_operational": 2, 00:13:44.768 "base_bdevs_list": [ 00:13:44.768 { 00:13:44.768 "name": null, 00:13:44.768 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:44.768 "is_configured": false, 00:13:44.768 "data_offset": 2048, 00:13:44.768 "data_size": 63488 00:13:44.768 }, 00:13:44.768 { 00:13:44.768 "name": "pt2", 00:13:44.768 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:44.768 "is_configured": true, 00:13:44.768 "data_offset": 2048, 00:13:44.768 "data_size": 63488 00:13:44.768 }, 00:13:44.768 { 00:13:44.768 "name": "pt3", 00:13:44.768 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:44.768 "is_configured": true, 00:13:44.768 "data_offset": 2048, 00:13:44.768 "data_size": 63488 00:13:44.768 } 00:13:44.768 ] 00:13:44.768 }' 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:44.768 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:45.028 [2024-12-05 19:04:02.430577] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:45.028 [2024-12-05 19:04:02.430603] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:45.028 [2024-12-05 19:04:02.430659] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:45.028 [2024-12-05 19:04:02.430719] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:45.028 [2024-12-05 19:04:02.430731] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:45.028 [2024-12-05 19:04:02.502447] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:45.028 [2024-12-05 19:04:02.502499] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:45.028 [2024-12-05 19:04:02.502511] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:13:45.028 [2024-12-05 19:04:02.502521] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:45.028 [2024-12-05 19:04:02.504507] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:45.028 [2024-12-05 19:04:02.504543] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:45.028 [2024-12-05 19:04:02.504597] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:13:45.028 [2024-12-05 19:04:02.504626] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:45.028 [2024-12-05 19:04:02.504716] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:13:45.028 [2024-12-05 19:04:02.504739] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:45.028 [2024-12-05 19:04:02.504758] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:13:45.028 [2024-12-05 19:04:02.504794] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:45.028 pt1 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:45.028 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:45.029 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.029 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:45.029 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.029 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:45.029 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.029 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:45.029 "name": "raid_bdev1", 00:13:45.029 "uuid": "364d237a-2043-4b9d-a0c0-2eca9c9bb6c8", 00:13:45.029 "strip_size_kb": 64, 00:13:45.029 "state": "configuring", 00:13:45.029 "raid_level": "raid5f", 00:13:45.029 "superblock": true, 00:13:45.029 "num_base_bdevs": 3, 00:13:45.029 "num_base_bdevs_discovered": 1, 00:13:45.029 "num_base_bdevs_operational": 2, 00:13:45.029 "base_bdevs_list": [ 00:13:45.029 { 00:13:45.029 "name": null, 00:13:45.029 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:45.029 "is_configured": false, 00:13:45.029 "data_offset": 2048, 00:13:45.029 "data_size": 63488 00:13:45.029 }, 00:13:45.029 { 00:13:45.029 "name": "pt2", 00:13:45.029 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:45.029 "is_configured": true, 00:13:45.029 "data_offset": 2048, 00:13:45.029 "data_size": 63488 00:13:45.029 }, 00:13:45.029 { 00:13:45.029 "name": null, 00:13:45.029 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:45.029 "is_configured": false, 00:13:45.029 "data_offset": 2048, 00:13:45.029 "data_size": 63488 00:13:45.029 } 00:13:45.029 ] 00:13:45.029 }' 00:13:45.029 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:45.029 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:45.600 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:13:45.600 19:04:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:13:45.600 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.600 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:45.600 19:04:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:45.600 [2024-12-05 19:04:03.013552] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:45.600 [2024-12-05 19:04:03.013648] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:45.600 [2024-12-05 19:04:03.013689] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:13:45.600 [2024-12-05 19:04:03.013716] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:45.600 [2024-12-05 19:04:03.014082] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:45.600 [2024-12-05 19:04:03.014147] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:45.600 [2024-12-05 19:04:03.014230] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:45.600 [2024-12-05 19:04:03.014281] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:45.600 [2024-12-05 19:04:03.014386] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:13:45.600 [2024-12-05 19:04:03.014440] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:45.600 [2024-12-05 19:04:03.014685] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:13:45.600 [2024-12-05 19:04:03.015154] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:13:45.600 [2024-12-05 19:04:03.015205] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:13:45.600 [2024-12-05 19:04:03.015393] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:45.600 pt3 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:45.600 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:45.600 "name": "raid_bdev1", 00:13:45.600 "uuid": "364d237a-2043-4b9d-a0c0-2eca9c9bb6c8", 00:13:45.600 "strip_size_kb": 64, 00:13:45.600 "state": "online", 00:13:45.600 "raid_level": "raid5f", 00:13:45.600 "superblock": true, 00:13:45.600 "num_base_bdevs": 3, 00:13:45.600 "num_base_bdevs_discovered": 2, 00:13:45.600 "num_base_bdevs_operational": 2, 00:13:45.600 "base_bdevs_list": [ 00:13:45.600 { 00:13:45.600 "name": null, 00:13:45.600 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:45.600 "is_configured": false, 00:13:45.600 "data_offset": 2048, 00:13:45.600 "data_size": 63488 00:13:45.600 }, 00:13:45.600 { 00:13:45.600 "name": "pt2", 00:13:45.601 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:45.601 "is_configured": true, 00:13:45.601 "data_offset": 2048, 00:13:45.601 "data_size": 63488 00:13:45.601 }, 00:13:45.601 { 00:13:45.601 "name": "pt3", 00:13:45.601 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:45.601 "is_configured": true, 00:13:45.601 "data_offset": 2048, 00:13:45.601 "data_size": 63488 00:13:45.601 } 00:13:45.601 ] 00:13:45.601 }' 00:13:45.601 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:45.601 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.171 [2024-12-05 19:04:03.512918] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 364d237a-2043-4b9d-a0c0-2eca9c9bb6c8 '!=' 364d237a-2043-4b9d-a0c0-2eca9c9bb6c8 ']' 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 89513 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 89513 ']' 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # kill -0 89513 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # uname 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 89513 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 89513' 00:13:46.171 killing process with pid 89513 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@973 -- # kill 89513 00:13:46.171 [2024-12-05 19:04:03.593527] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:46.171 [2024-12-05 19:04:03.593585] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:46.171 [2024-12-05 19:04:03.593634] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:46.171 [2024-12-05 19:04:03.593643] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:13:46.171 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@978 -- # wait 89513 00:13:46.171 [2024-12-05 19:04:03.626308] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:46.432 19:04:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:13:46.432 00:13:46.432 real 0m6.412s 00:13:46.432 user 0m10.732s 00:13:46.432 sys 0m1.428s 00:13:46.432 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:46.432 ************************************ 00:13:46.432 END TEST raid5f_superblock_test 00:13:46.432 ************************************ 00:13:46.432 19:04:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.432 19:04:03 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:13:46.432 19:04:03 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 3 false false true 00:13:46.432 19:04:03 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:13:46.432 19:04:03 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:46.432 19:04:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:46.432 ************************************ 00:13:46.432 START TEST raid5f_rebuild_test 00:13:46.432 ************************************ 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 3 false false true 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=89904 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 89904 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 89904 ']' 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:46.432 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:46.432 19:04:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:46.692 [2024-12-05 19:04:04.019131] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:13:46.692 [2024-12-05 19:04:04.019311] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:13:46.692 Zero copy mechanism will not be used. 00:13:46.692 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89904 ] 00:13:46.692 [2024-12-05 19:04:04.173979] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:46.692 [2024-12-05 19:04:04.199485] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:46.692 [2024-12-05 19:04:04.243050] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:46.692 [2024-12-05 19:04:04.243159] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.632 BaseBdev1_malloc 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.632 [2024-12-05 19:04:04.879041] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:47.632 [2024-12-05 19:04:04.879106] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:47.632 [2024-12-05 19:04:04.879131] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:47.632 [2024-12-05 19:04:04.879142] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:47.632 [2024-12-05 19:04:04.881242] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:47.632 [2024-12-05 19:04:04.881278] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:47.632 BaseBdev1 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.632 BaseBdev2_malloc 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.632 [2024-12-05 19:04:04.907612] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:47.632 [2024-12-05 19:04:04.907683] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:47.632 [2024-12-05 19:04:04.907707] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:47.632 [2024-12-05 19:04:04.907715] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:47.632 [2024-12-05 19:04:04.909787] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:47.632 [2024-12-05 19:04:04.909826] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:47.632 BaseBdev2 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.632 BaseBdev3_malloc 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.632 [2024-12-05 19:04:04.936134] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:47.632 [2024-12-05 19:04:04.936186] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:47.632 [2024-12-05 19:04:04.936209] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:47.632 [2024-12-05 19:04:04.936217] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:47.632 [2024-12-05 19:04:04.938251] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:47.632 [2024-12-05 19:04:04.938287] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:47.632 BaseBdev3 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.632 spare_malloc 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.632 spare_delay 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.632 19:04:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.632 [2024-12-05 19:04:04.996940] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:47.632 [2024-12-05 19:04:04.997016] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:47.632 [2024-12-05 19:04:04.997057] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:13:47.632 [2024-12-05 19:04:04.997073] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:47.632 [2024-12-05 19:04:04.999623] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:47.632 [2024-12-05 19:04:04.999768] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:47.632 spare 00:13:47.632 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.633 [2024-12-05 19:04:05.008953] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:47.633 [2024-12-05 19:04:05.010792] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:47.633 [2024-12-05 19:04:05.010855] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:47.633 [2024-12-05 19:04:05.010932] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:47.633 [2024-12-05 19:04:05.010943] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:47.633 [2024-12-05 19:04:05.011213] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:13:47.633 [2024-12-05 19:04:05.011629] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:47.633 [2024-12-05 19:04:05.011640] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:47.633 [2024-12-05 19:04:05.011771] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:47.633 "name": "raid_bdev1", 00:13:47.633 "uuid": "bf2a5525-3446-49cb-baf8-7becde1ee151", 00:13:47.633 "strip_size_kb": 64, 00:13:47.633 "state": "online", 00:13:47.633 "raid_level": "raid5f", 00:13:47.633 "superblock": false, 00:13:47.633 "num_base_bdevs": 3, 00:13:47.633 "num_base_bdevs_discovered": 3, 00:13:47.633 "num_base_bdevs_operational": 3, 00:13:47.633 "base_bdevs_list": [ 00:13:47.633 { 00:13:47.633 "name": "BaseBdev1", 00:13:47.633 "uuid": "6d54d937-13c8-5d18-9244-4c186d511fe5", 00:13:47.633 "is_configured": true, 00:13:47.633 "data_offset": 0, 00:13:47.633 "data_size": 65536 00:13:47.633 }, 00:13:47.633 { 00:13:47.633 "name": "BaseBdev2", 00:13:47.633 "uuid": "db227b61-b38e-550b-897a-566993a31ba3", 00:13:47.633 "is_configured": true, 00:13:47.633 "data_offset": 0, 00:13:47.633 "data_size": 65536 00:13:47.633 }, 00:13:47.633 { 00:13:47.633 "name": "BaseBdev3", 00:13:47.633 "uuid": "b1fd5b15-dd53-5fea-8750-26f83eebbe22", 00:13:47.633 "is_configured": true, 00:13:47.633 "data_offset": 0, 00:13:47.633 "data_size": 65536 00:13:47.633 } 00:13:47.633 ] 00:13:47.633 }' 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:47.633 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.200 [2024-12-05 19:04:05.492699] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=131072 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:48.200 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:13:48.459 [2024-12-05 19:04:05.768096] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:13:48.460 /dev/nbd0 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:48.460 1+0 records in 00:13:48.460 1+0 records out 00:13:48.460 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000477274 s, 8.6 MB/s 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 128 00:13:48.460 19:04:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=512 oflag=direct 00:13:48.718 512+0 records in 00:13:48.718 512+0 records out 00:13:48.718 67108864 bytes (67 MB, 64 MiB) copied, 0.279385 s, 240 MB/s 00:13:48.718 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:48.718 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:48.718 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:48.718 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:48.718 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:13:48.718 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:48.718 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:48.977 [2024-12-05 19:04:06.342151] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.977 [2024-12-05 19:04:06.355440] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:48.977 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:48.977 "name": "raid_bdev1", 00:13:48.977 "uuid": "bf2a5525-3446-49cb-baf8-7becde1ee151", 00:13:48.977 "strip_size_kb": 64, 00:13:48.977 "state": "online", 00:13:48.977 "raid_level": "raid5f", 00:13:48.978 "superblock": false, 00:13:48.978 "num_base_bdevs": 3, 00:13:48.978 "num_base_bdevs_discovered": 2, 00:13:48.978 "num_base_bdevs_operational": 2, 00:13:48.978 "base_bdevs_list": [ 00:13:48.978 { 00:13:48.978 "name": null, 00:13:48.978 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:48.978 "is_configured": false, 00:13:48.978 "data_offset": 0, 00:13:48.978 "data_size": 65536 00:13:48.978 }, 00:13:48.978 { 00:13:48.978 "name": "BaseBdev2", 00:13:48.978 "uuid": "db227b61-b38e-550b-897a-566993a31ba3", 00:13:48.978 "is_configured": true, 00:13:48.978 "data_offset": 0, 00:13:48.978 "data_size": 65536 00:13:48.978 }, 00:13:48.978 { 00:13:48.978 "name": "BaseBdev3", 00:13:48.978 "uuid": "b1fd5b15-dd53-5fea-8750-26f83eebbe22", 00:13:48.978 "is_configured": true, 00:13:48.978 "data_offset": 0, 00:13:48.978 "data_size": 65536 00:13:48.978 } 00:13:48.978 ] 00:13:48.978 }' 00:13:48.978 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:48.978 19:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:49.236 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:49.236 19:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:49.236 19:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:49.236 [2024-12-05 19:04:06.766734] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:49.236 [2024-12-05 19:04:06.771323] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027cd0 00:13:49.236 19:04:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:49.236 19:04:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:49.236 [2024-12-05 19:04:06.773534] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:50.613 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:50.613 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:50.613 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:50.613 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:50.613 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:50.613 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.613 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:50.613 19:04:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.613 19:04:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.613 19:04:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.613 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:50.613 "name": "raid_bdev1", 00:13:50.613 "uuid": "bf2a5525-3446-49cb-baf8-7becde1ee151", 00:13:50.613 "strip_size_kb": 64, 00:13:50.613 "state": "online", 00:13:50.613 "raid_level": "raid5f", 00:13:50.613 "superblock": false, 00:13:50.613 "num_base_bdevs": 3, 00:13:50.613 "num_base_bdevs_discovered": 3, 00:13:50.613 "num_base_bdevs_operational": 3, 00:13:50.614 "process": { 00:13:50.614 "type": "rebuild", 00:13:50.614 "target": "spare", 00:13:50.614 "progress": { 00:13:50.614 "blocks": 20480, 00:13:50.614 "percent": 15 00:13:50.614 } 00:13:50.614 }, 00:13:50.614 "base_bdevs_list": [ 00:13:50.614 { 00:13:50.614 "name": "spare", 00:13:50.614 "uuid": "0d02025f-908f-5761-b5cb-a6652545dc12", 00:13:50.614 "is_configured": true, 00:13:50.614 "data_offset": 0, 00:13:50.614 "data_size": 65536 00:13:50.614 }, 00:13:50.614 { 00:13:50.614 "name": "BaseBdev2", 00:13:50.614 "uuid": "db227b61-b38e-550b-897a-566993a31ba3", 00:13:50.614 "is_configured": true, 00:13:50.614 "data_offset": 0, 00:13:50.614 "data_size": 65536 00:13:50.614 }, 00:13:50.614 { 00:13:50.614 "name": "BaseBdev3", 00:13:50.614 "uuid": "b1fd5b15-dd53-5fea-8750-26f83eebbe22", 00:13:50.614 "is_configured": true, 00:13:50.614 "data_offset": 0, 00:13:50.614 "data_size": 65536 00:13:50.614 } 00:13:50.614 ] 00:13:50.614 }' 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.614 [2024-12-05 19:04:07.909638] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:50.614 [2024-12-05 19:04:07.980272] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:50.614 [2024-12-05 19:04:07.980385] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:50.614 [2024-12-05 19:04:07.980421] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:50.614 [2024-12-05 19:04:07.980460] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.614 19:04:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:50.614 19:04:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:50.614 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:50.614 "name": "raid_bdev1", 00:13:50.614 "uuid": "bf2a5525-3446-49cb-baf8-7becde1ee151", 00:13:50.614 "strip_size_kb": 64, 00:13:50.614 "state": "online", 00:13:50.614 "raid_level": "raid5f", 00:13:50.614 "superblock": false, 00:13:50.614 "num_base_bdevs": 3, 00:13:50.614 "num_base_bdevs_discovered": 2, 00:13:50.614 "num_base_bdevs_operational": 2, 00:13:50.614 "base_bdevs_list": [ 00:13:50.614 { 00:13:50.614 "name": null, 00:13:50.614 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:50.614 "is_configured": false, 00:13:50.614 "data_offset": 0, 00:13:50.614 "data_size": 65536 00:13:50.614 }, 00:13:50.614 { 00:13:50.614 "name": "BaseBdev2", 00:13:50.614 "uuid": "db227b61-b38e-550b-897a-566993a31ba3", 00:13:50.614 "is_configured": true, 00:13:50.614 "data_offset": 0, 00:13:50.614 "data_size": 65536 00:13:50.614 }, 00:13:50.614 { 00:13:50.614 "name": "BaseBdev3", 00:13:50.614 "uuid": "b1fd5b15-dd53-5fea-8750-26f83eebbe22", 00:13:50.614 "is_configured": true, 00:13:50.614 "data_offset": 0, 00:13:50.614 "data_size": 65536 00:13:50.614 } 00:13:50.614 ] 00:13:50.614 }' 00:13:50.614 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:50.614 19:04:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.873 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:50.873 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:50.873 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:50.873 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:51.133 "name": "raid_bdev1", 00:13:51.133 "uuid": "bf2a5525-3446-49cb-baf8-7becde1ee151", 00:13:51.133 "strip_size_kb": 64, 00:13:51.133 "state": "online", 00:13:51.133 "raid_level": "raid5f", 00:13:51.133 "superblock": false, 00:13:51.133 "num_base_bdevs": 3, 00:13:51.133 "num_base_bdevs_discovered": 2, 00:13:51.133 "num_base_bdevs_operational": 2, 00:13:51.133 "base_bdevs_list": [ 00:13:51.133 { 00:13:51.133 "name": null, 00:13:51.133 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:51.133 "is_configured": false, 00:13:51.133 "data_offset": 0, 00:13:51.133 "data_size": 65536 00:13:51.133 }, 00:13:51.133 { 00:13:51.133 "name": "BaseBdev2", 00:13:51.133 "uuid": "db227b61-b38e-550b-897a-566993a31ba3", 00:13:51.133 "is_configured": true, 00:13:51.133 "data_offset": 0, 00:13:51.133 "data_size": 65536 00:13:51.133 }, 00:13:51.133 { 00:13:51.133 "name": "BaseBdev3", 00:13:51.133 "uuid": "b1fd5b15-dd53-5fea-8750-26f83eebbe22", 00:13:51.133 "is_configured": true, 00:13:51.133 "data_offset": 0, 00:13:51.133 "data_size": 65536 00:13:51.133 } 00:13:51.133 ] 00:13:51.133 }' 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.133 [2024-12-05 19:04:08.589207] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:51.133 [2024-12-05 19:04:08.592980] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027da0 00:13:51.133 [2024-12-05 19:04:08.595142] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:51.133 19:04:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:52.074 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:52.074 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:52.074 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:52.074 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:52.074 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:52.074 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.074 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:52.074 19:04:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.074 19:04:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.074 19:04:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.333 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:52.333 "name": "raid_bdev1", 00:13:52.333 "uuid": "bf2a5525-3446-49cb-baf8-7becde1ee151", 00:13:52.333 "strip_size_kb": 64, 00:13:52.334 "state": "online", 00:13:52.334 "raid_level": "raid5f", 00:13:52.334 "superblock": false, 00:13:52.334 "num_base_bdevs": 3, 00:13:52.334 "num_base_bdevs_discovered": 3, 00:13:52.334 "num_base_bdevs_operational": 3, 00:13:52.334 "process": { 00:13:52.334 "type": "rebuild", 00:13:52.334 "target": "spare", 00:13:52.334 "progress": { 00:13:52.334 "blocks": 20480, 00:13:52.334 "percent": 15 00:13:52.334 } 00:13:52.334 }, 00:13:52.334 "base_bdevs_list": [ 00:13:52.334 { 00:13:52.334 "name": "spare", 00:13:52.334 "uuid": "0d02025f-908f-5761-b5cb-a6652545dc12", 00:13:52.334 "is_configured": true, 00:13:52.334 "data_offset": 0, 00:13:52.334 "data_size": 65536 00:13:52.334 }, 00:13:52.334 { 00:13:52.334 "name": "BaseBdev2", 00:13:52.334 "uuid": "db227b61-b38e-550b-897a-566993a31ba3", 00:13:52.334 "is_configured": true, 00:13:52.334 "data_offset": 0, 00:13:52.334 "data_size": 65536 00:13:52.334 }, 00:13:52.334 { 00:13:52.334 "name": "BaseBdev3", 00:13:52.334 "uuid": "b1fd5b15-dd53-5fea-8750-26f83eebbe22", 00:13:52.334 "is_configured": true, 00:13:52.334 "data_offset": 0, 00:13:52.334 "data_size": 65536 00:13:52.334 } 00:13:52.334 ] 00:13:52.334 }' 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=439 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:52.334 "name": "raid_bdev1", 00:13:52.334 "uuid": "bf2a5525-3446-49cb-baf8-7becde1ee151", 00:13:52.334 "strip_size_kb": 64, 00:13:52.334 "state": "online", 00:13:52.334 "raid_level": "raid5f", 00:13:52.334 "superblock": false, 00:13:52.334 "num_base_bdevs": 3, 00:13:52.334 "num_base_bdevs_discovered": 3, 00:13:52.334 "num_base_bdevs_operational": 3, 00:13:52.334 "process": { 00:13:52.334 "type": "rebuild", 00:13:52.334 "target": "spare", 00:13:52.334 "progress": { 00:13:52.334 "blocks": 22528, 00:13:52.334 "percent": 17 00:13:52.334 } 00:13:52.334 }, 00:13:52.334 "base_bdevs_list": [ 00:13:52.334 { 00:13:52.334 "name": "spare", 00:13:52.334 "uuid": "0d02025f-908f-5761-b5cb-a6652545dc12", 00:13:52.334 "is_configured": true, 00:13:52.334 "data_offset": 0, 00:13:52.334 "data_size": 65536 00:13:52.334 }, 00:13:52.334 { 00:13:52.334 "name": "BaseBdev2", 00:13:52.334 "uuid": "db227b61-b38e-550b-897a-566993a31ba3", 00:13:52.334 "is_configured": true, 00:13:52.334 "data_offset": 0, 00:13:52.334 "data_size": 65536 00:13:52.334 }, 00:13:52.334 { 00:13:52.334 "name": "BaseBdev3", 00:13:52.334 "uuid": "b1fd5b15-dd53-5fea-8750-26f83eebbe22", 00:13:52.334 "is_configured": true, 00:13:52.334 "data_offset": 0, 00:13:52.334 "data_size": 65536 00:13:52.334 } 00:13:52.334 ] 00:13:52.334 }' 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:52.334 19:04:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:53.715 19:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:53.715 19:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:53.715 19:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:53.715 19:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:53.715 19:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:53.715 19:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:53.715 19:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.715 19:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:53.715 19:04:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:53.715 19:04:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.715 19:04:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:53.715 19:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:53.715 "name": "raid_bdev1", 00:13:53.715 "uuid": "bf2a5525-3446-49cb-baf8-7becde1ee151", 00:13:53.715 "strip_size_kb": 64, 00:13:53.715 "state": "online", 00:13:53.715 "raid_level": "raid5f", 00:13:53.715 "superblock": false, 00:13:53.715 "num_base_bdevs": 3, 00:13:53.715 "num_base_bdevs_discovered": 3, 00:13:53.715 "num_base_bdevs_operational": 3, 00:13:53.715 "process": { 00:13:53.715 "type": "rebuild", 00:13:53.715 "target": "spare", 00:13:53.715 "progress": { 00:13:53.715 "blocks": 45056, 00:13:53.715 "percent": 34 00:13:53.715 } 00:13:53.715 }, 00:13:53.715 "base_bdevs_list": [ 00:13:53.715 { 00:13:53.715 "name": "spare", 00:13:53.715 "uuid": "0d02025f-908f-5761-b5cb-a6652545dc12", 00:13:53.715 "is_configured": true, 00:13:53.715 "data_offset": 0, 00:13:53.715 "data_size": 65536 00:13:53.715 }, 00:13:53.715 { 00:13:53.715 "name": "BaseBdev2", 00:13:53.715 "uuid": "db227b61-b38e-550b-897a-566993a31ba3", 00:13:53.715 "is_configured": true, 00:13:53.715 "data_offset": 0, 00:13:53.715 "data_size": 65536 00:13:53.715 }, 00:13:53.715 { 00:13:53.715 "name": "BaseBdev3", 00:13:53.716 "uuid": "b1fd5b15-dd53-5fea-8750-26f83eebbe22", 00:13:53.716 "is_configured": true, 00:13:53.716 "data_offset": 0, 00:13:53.716 "data_size": 65536 00:13:53.716 } 00:13:53.716 ] 00:13:53.716 }' 00:13:53.716 19:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:53.716 19:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:53.716 19:04:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:53.716 19:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:53.716 19:04:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:54.656 "name": "raid_bdev1", 00:13:54.656 "uuid": "bf2a5525-3446-49cb-baf8-7becde1ee151", 00:13:54.656 "strip_size_kb": 64, 00:13:54.656 "state": "online", 00:13:54.656 "raid_level": "raid5f", 00:13:54.656 "superblock": false, 00:13:54.656 "num_base_bdevs": 3, 00:13:54.656 "num_base_bdevs_discovered": 3, 00:13:54.656 "num_base_bdevs_operational": 3, 00:13:54.656 "process": { 00:13:54.656 "type": "rebuild", 00:13:54.656 "target": "spare", 00:13:54.656 "progress": { 00:13:54.656 "blocks": 69632, 00:13:54.656 "percent": 53 00:13:54.656 } 00:13:54.656 }, 00:13:54.656 "base_bdevs_list": [ 00:13:54.656 { 00:13:54.656 "name": "spare", 00:13:54.656 "uuid": "0d02025f-908f-5761-b5cb-a6652545dc12", 00:13:54.656 "is_configured": true, 00:13:54.656 "data_offset": 0, 00:13:54.656 "data_size": 65536 00:13:54.656 }, 00:13:54.656 { 00:13:54.656 "name": "BaseBdev2", 00:13:54.656 "uuid": "db227b61-b38e-550b-897a-566993a31ba3", 00:13:54.656 "is_configured": true, 00:13:54.656 "data_offset": 0, 00:13:54.656 "data_size": 65536 00:13:54.656 }, 00:13:54.656 { 00:13:54.656 "name": "BaseBdev3", 00:13:54.656 "uuid": "b1fd5b15-dd53-5fea-8750-26f83eebbe22", 00:13:54.656 "is_configured": true, 00:13:54.656 "data_offset": 0, 00:13:54.656 "data_size": 65536 00:13:54.656 } 00:13:54.656 ] 00:13:54.656 }' 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:54.656 19:04:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:56.068 19:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:56.068 19:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:56.068 19:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:56.068 19:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:56.068 19:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:56.068 19:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:56.068 19:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.068 19:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:56.068 19:04:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:56.068 19:04:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.068 19:04:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:56.068 19:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:56.068 "name": "raid_bdev1", 00:13:56.068 "uuid": "bf2a5525-3446-49cb-baf8-7becde1ee151", 00:13:56.068 "strip_size_kb": 64, 00:13:56.068 "state": "online", 00:13:56.068 "raid_level": "raid5f", 00:13:56.068 "superblock": false, 00:13:56.068 "num_base_bdevs": 3, 00:13:56.068 "num_base_bdevs_discovered": 3, 00:13:56.068 "num_base_bdevs_operational": 3, 00:13:56.068 "process": { 00:13:56.068 "type": "rebuild", 00:13:56.068 "target": "spare", 00:13:56.068 "progress": { 00:13:56.068 "blocks": 92160, 00:13:56.068 "percent": 70 00:13:56.068 } 00:13:56.068 }, 00:13:56.068 "base_bdevs_list": [ 00:13:56.068 { 00:13:56.068 "name": "spare", 00:13:56.068 "uuid": "0d02025f-908f-5761-b5cb-a6652545dc12", 00:13:56.068 "is_configured": true, 00:13:56.068 "data_offset": 0, 00:13:56.068 "data_size": 65536 00:13:56.068 }, 00:13:56.068 { 00:13:56.068 "name": "BaseBdev2", 00:13:56.068 "uuid": "db227b61-b38e-550b-897a-566993a31ba3", 00:13:56.068 "is_configured": true, 00:13:56.068 "data_offset": 0, 00:13:56.068 "data_size": 65536 00:13:56.068 }, 00:13:56.068 { 00:13:56.068 "name": "BaseBdev3", 00:13:56.068 "uuid": "b1fd5b15-dd53-5fea-8750-26f83eebbe22", 00:13:56.068 "is_configured": true, 00:13:56.068 "data_offset": 0, 00:13:56.068 "data_size": 65536 00:13:56.068 } 00:13:56.068 ] 00:13:56.069 }' 00:13:56.069 19:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:56.069 19:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:56.069 19:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:56.069 19:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:56.069 19:04:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:57.020 "name": "raid_bdev1", 00:13:57.020 "uuid": "bf2a5525-3446-49cb-baf8-7becde1ee151", 00:13:57.020 "strip_size_kb": 64, 00:13:57.020 "state": "online", 00:13:57.020 "raid_level": "raid5f", 00:13:57.020 "superblock": false, 00:13:57.020 "num_base_bdevs": 3, 00:13:57.020 "num_base_bdevs_discovered": 3, 00:13:57.020 "num_base_bdevs_operational": 3, 00:13:57.020 "process": { 00:13:57.020 "type": "rebuild", 00:13:57.020 "target": "spare", 00:13:57.020 "progress": { 00:13:57.020 "blocks": 116736, 00:13:57.020 "percent": 89 00:13:57.020 } 00:13:57.020 }, 00:13:57.020 "base_bdevs_list": [ 00:13:57.020 { 00:13:57.020 "name": "spare", 00:13:57.020 "uuid": "0d02025f-908f-5761-b5cb-a6652545dc12", 00:13:57.020 "is_configured": true, 00:13:57.020 "data_offset": 0, 00:13:57.020 "data_size": 65536 00:13:57.020 }, 00:13:57.020 { 00:13:57.020 "name": "BaseBdev2", 00:13:57.020 "uuid": "db227b61-b38e-550b-897a-566993a31ba3", 00:13:57.020 "is_configured": true, 00:13:57.020 "data_offset": 0, 00:13:57.020 "data_size": 65536 00:13:57.020 }, 00:13:57.020 { 00:13:57.020 "name": "BaseBdev3", 00:13:57.020 "uuid": "b1fd5b15-dd53-5fea-8750-26f83eebbe22", 00:13:57.020 "is_configured": true, 00:13:57.020 "data_offset": 0, 00:13:57.020 "data_size": 65536 00:13:57.020 } 00:13:57.020 ] 00:13:57.020 }' 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:57.020 19:04:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:57.591 [2024-12-05 19:04:15.027694] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:57.591 [2024-12-05 19:04:15.027759] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:57.591 [2024-12-05 19:04:15.027808] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:58.158 "name": "raid_bdev1", 00:13:58.158 "uuid": "bf2a5525-3446-49cb-baf8-7becde1ee151", 00:13:58.158 "strip_size_kb": 64, 00:13:58.158 "state": "online", 00:13:58.158 "raid_level": "raid5f", 00:13:58.158 "superblock": false, 00:13:58.158 "num_base_bdevs": 3, 00:13:58.158 "num_base_bdevs_discovered": 3, 00:13:58.158 "num_base_bdevs_operational": 3, 00:13:58.158 "base_bdevs_list": [ 00:13:58.158 { 00:13:58.158 "name": "spare", 00:13:58.158 "uuid": "0d02025f-908f-5761-b5cb-a6652545dc12", 00:13:58.158 "is_configured": true, 00:13:58.158 "data_offset": 0, 00:13:58.158 "data_size": 65536 00:13:58.158 }, 00:13:58.158 { 00:13:58.158 "name": "BaseBdev2", 00:13:58.158 "uuid": "db227b61-b38e-550b-897a-566993a31ba3", 00:13:58.158 "is_configured": true, 00:13:58.158 "data_offset": 0, 00:13:58.158 "data_size": 65536 00:13:58.158 }, 00:13:58.158 { 00:13:58.158 "name": "BaseBdev3", 00:13:58.158 "uuid": "b1fd5b15-dd53-5fea-8750-26f83eebbe22", 00:13:58.158 "is_configured": true, 00:13:58.158 "data_offset": 0, 00:13:58.158 "data_size": 65536 00:13:58.158 } 00:13:58.158 ] 00:13:58.158 }' 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:58.158 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:58.158 "name": "raid_bdev1", 00:13:58.158 "uuid": "bf2a5525-3446-49cb-baf8-7becde1ee151", 00:13:58.158 "strip_size_kb": 64, 00:13:58.158 "state": "online", 00:13:58.158 "raid_level": "raid5f", 00:13:58.158 "superblock": false, 00:13:58.158 "num_base_bdevs": 3, 00:13:58.158 "num_base_bdevs_discovered": 3, 00:13:58.158 "num_base_bdevs_operational": 3, 00:13:58.158 "base_bdevs_list": [ 00:13:58.158 { 00:13:58.158 "name": "spare", 00:13:58.158 "uuid": "0d02025f-908f-5761-b5cb-a6652545dc12", 00:13:58.158 "is_configured": true, 00:13:58.159 "data_offset": 0, 00:13:58.159 "data_size": 65536 00:13:58.159 }, 00:13:58.159 { 00:13:58.159 "name": "BaseBdev2", 00:13:58.159 "uuid": "db227b61-b38e-550b-897a-566993a31ba3", 00:13:58.159 "is_configured": true, 00:13:58.159 "data_offset": 0, 00:13:58.159 "data_size": 65536 00:13:58.159 }, 00:13:58.159 { 00:13:58.159 "name": "BaseBdev3", 00:13:58.159 "uuid": "b1fd5b15-dd53-5fea-8750-26f83eebbe22", 00:13:58.159 "is_configured": true, 00:13:58.159 "data_offset": 0, 00:13:58.159 "data_size": 65536 00:13:58.159 } 00:13:58.159 ] 00:13:58.159 }' 00:13:58.159 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:58.159 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:58.159 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:58.417 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:58.417 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:58.417 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:58.417 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:58.417 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:58.418 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:58.418 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:58.418 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:58.418 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:58.418 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:58.418 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:58.418 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:58.418 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:58.418 19:04:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:58.418 19:04:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:58.418 19:04:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:58.418 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:58.418 "name": "raid_bdev1", 00:13:58.418 "uuid": "bf2a5525-3446-49cb-baf8-7becde1ee151", 00:13:58.418 "strip_size_kb": 64, 00:13:58.418 "state": "online", 00:13:58.418 "raid_level": "raid5f", 00:13:58.418 "superblock": false, 00:13:58.418 "num_base_bdevs": 3, 00:13:58.418 "num_base_bdevs_discovered": 3, 00:13:58.418 "num_base_bdevs_operational": 3, 00:13:58.418 "base_bdevs_list": [ 00:13:58.418 { 00:13:58.418 "name": "spare", 00:13:58.418 "uuid": "0d02025f-908f-5761-b5cb-a6652545dc12", 00:13:58.418 "is_configured": true, 00:13:58.418 "data_offset": 0, 00:13:58.418 "data_size": 65536 00:13:58.418 }, 00:13:58.418 { 00:13:58.418 "name": "BaseBdev2", 00:13:58.418 "uuid": "db227b61-b38e-550b-897a-566993a31ba3", 00:13:58.418 "is_configured": true, 00:13:58.418 "data_offset": 0, 00:13:58.418 "data_size": 65536 00:13:58.418 }, 00:13:58.418 { 00:13:58.418 "name": "BaseBdev3", 00:13:58.418 "uuid": "b1fd5b15-dd53-5fea-8750-26f83eebbe22", 00:13:58.418 "is_configured": true, 00:13:58.418 "data_offset": 0, 00:13:58.418 "data_size": 65536 00:13:58.418 } 00:13:58.418 ] 00:13:58.418 }' 00:13:58.418 19:04:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:58.418 19:04:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:58.677 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:58.677 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:58.677 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:58.677 [2024-12-05 19:04:16.187199] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:58.677 [2024-12-05 19:04:16.187283] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:58.677 [2024-12-05 19:04:16.187401] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:58.677 [2024-12-05 19:04:16.187504] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:58.677 [2024-12-05 19:04:16.187562] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:58.677 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:58.677 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:58.677 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:13:58.677 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:58.677 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:58.677 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:58.937 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:58.937 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:58.937 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:13:58.937 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:13:58.937 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:58.937 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:13:58.937 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:58.937 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:58.937 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:58.937 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:13:58.937 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:58.937 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:58.937 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:13:58.937 /dev/nbd0 00:13:58.937 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:58.937 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:58.938 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:13:58.938 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:13:58.938 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:58.938 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:58.938 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:13:58.938 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:13:58.938 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:58.938 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:58.938 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:58.938 1+0 records in 00:13:58.938 1+0 records out 00:13:58.938 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000564668 s, 7.3 MB/s 00:13:58.938 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:13:59.199 /dev/nbd1 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:59.199 1+0 records in 00:13:59.199 1+0 records out 00:13:59.199 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000373075 s, 11.0 MB/s 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:59.199 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:59.459 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:13:59.460 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:59.460 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:59.460 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:59.460 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:13:59.460 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:59.460 19:04:16 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 89904 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 89904 ']' 00:13:59.720 19:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 89904 00:13:59.981 19:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:13:59.981 19:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:59.981 19:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 89904 00:13:59.981 19:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:59.981 19:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:59.981 killing process with pid 89904 00:13:59.981 19:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 89904' 00:13:59.981 19:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@973 -- # kill 89904 00:13:59.981 Received shutdown signal, test time was about 60.000000 seconds 00:13:59.981 00:13:59.981 Latency(us) 00:13:59.981 [2024-12-05T19:04:17.540Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:59.981 [2024-12-05T19:04:17.540Z] =================================================================================================================== 00:13:59.981 [2024-12-05T19:04:17.540Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:13:59.981 [2024-12-05 19:04:17.321122] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:59.981 19:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@978 -- # wait 89904 00:13:59.981 [2024-12-05 19:04:17.360860] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:00.241 19:04:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:14:00.241 00:14:00.241 real 0m13.634s 00:14:00.241 user 0m17.013s 00:14:00.241 sys 0m2.015s 00:14:00.241 19:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:00.241 19:04:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:00.241 ************************************ 00:14:00.241 END TEST raid5f_rebuild_test 00:14:00.241 ************************************ 00:14:00.241 19:04:17 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 3 true false true 00:14:00.241 19:04:17 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:14:00.241 19:04:17 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:00.241 19:04:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:00.241 ************************************ 00:14:00.241 START TEST raid5f_rebuild_test_sb 00:14:00.241 ************************************ 00:14:00.241 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 3 true false true 00:14:00.241 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:14:00.241 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:14:00.241 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:14:00.241 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:00.241 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:00.241 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:00.241 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:00.241 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:00.241 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:00.241 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=90244 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 90244 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 90244 ']' 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:00.242 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:00.242 19:04:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.242 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:00.242 Zero copy mechanism will not be used. 00:14:00.242 [2024-12-05 19:04:17.729883] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:14:00.242 [2024-12-05 19:04:17.730005] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid90244 ] 00:14:00.501 [2024-12-05 19:04:17.885441] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:00.501 [2024-12-05 19:04:17.910310] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:00.501 [2024-12-05 19:04:17.953780] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:00.501 [2024-12-05 19:04:17.953816] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.071 BaseBdev1_malloc 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.071 [2024-12-05 19:04:18.581782] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:01.071 [2024-12-05 19:04:18.581844] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:01.071 [2024-12-05 19:04:18.581870] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:01.071 [2024-12-05 19:04:18.581883] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:01.071 [2024-12-05 19:04:18.584005] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:01.071 [2024-12-05 19:04:18.584038] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:01.071 BaseBdev1 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.071 BaseBdev2_malloc 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.071 [2024-12-05 19:04:18.610328] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:01.071 [2024-12-05 19:04:18.610381] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:01.071 [2024-12-05 19:04:18.610402] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:01.071 [2024-12-05 19:04:18.610410] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:01.071 [2024-12-05 19:04:18.612469] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:01.071 [2024-12-05 19:04:18.612509] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:01.071 BaseBdev2 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.071 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.331 BaseBdev3_malloc 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.331 [2024-12-05 19:04:18.638921] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:01.331 [2024-12-05 19:04:18.638972] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:01.331 [2024-12-05 19:04:18.638994] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:01.331 [2024-12-05 19:04:18.639002] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:01.331 [2024-12-05 19:04:18.641055] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:01.331 [2024-12-05 19:04:18.641090] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:01.331 BaseBdev3 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.331 spare_malloc 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.331 spare_delay 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.331 [2024-12-05 19:04:18.687212] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:01.331 [2024-12-05 19:04:18.687265] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:01.331 [2024-12-05 19:04:18.687289] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:14:01.331 [2024-12-05 19:04:18.687297] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:01.331 [2024-12-05 19:04:18.689345] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:01.331 [2024-12-05 19:04:18.689380] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:01.331 spare 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.331 [2024-12-05 19:04:18.699253] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:01.331 [2024-12-05 19:04:18.701048] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:01.331 [2024-12-05 19:04:18.701108] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:01.331 [2024-12-05 19:04:18.701267] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:01.331 [2024-12-05 19:04:18.701284] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:01.331 [2024-12-05 19:04:18.701549] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:01.331 [2024-12-05 19:04:18.701970] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:01.331 [2024-12-05 19:04:18.701995] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:01.331 [2024-12-05 19:04:18.702110] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:01.331 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:01.332 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:01.332 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:01.332 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:01.332 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:01.332 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:01.332 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.332 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:01.332 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.332 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.332 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.332 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:01.332 "name": "raid_bdev1", 00:14:01.332 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:01.332 "strip_size_kb": 64, 00:14:01.332 "state": "online", 00:14:01.332 "raid_level": "raid5f", 00:14:01.332 "superblock": true, 00:14:01.332 "num_base_bdevs": 3, 00:14:01.332 "num_base_bdevs_discovered": 3, 00:14:01.332 "num_base_bdevs_operational": 3, 00:14:01.332 "base_bdevs_list": [ 00:14:01.332 { 00:14:01.332 "name": "BaseBdev1", 00:14:01.332 "uuid": "2d3fc5dd-b387-53fd-b330-a1cf8a4c2988", 00:14:01.332 "is_configured": true, 00:14:01.332 "data_offset": 2048, 00:14:01.332 "data_size": 63488 00:14:01.332 }, 00:14:01.332 { 00:14:01.332 "name": "BaseBdev2", 00:14:01.332 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:01.332 "is_configured": true, 00:14:01.332 "data_offset": 2048, 00:14:01.332 "data_size": 63488 00:14:01.332 }, 00:14:01.332 { 00:14:01.332 "name": "BaseBdev3", 00:14:01.332 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:01.332 "is_configured": true, 00:14:01.332 "data_offset": 2048, 00:14:01.332 "data_size": 63488 00:14:01.332 } 00:14:01.332 ] 00:14:01.332 }' 00:14:01.332 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:01.332 19:04:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.592 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:01.592 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:01.592 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.592 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.851 [2024-12-05 19:04:19.150875] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:01.851 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.851 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=126976 00:14:01.851 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:01.852 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:01.852 [2024-12-05 19:04:19.406267] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:14:02.112 /dev/nbd0 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:02.112 1+0 records in 00:14:02.112 1+0 records out 00:14:02.112 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00041318 s, 9.9 MB/s 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 128 00:14:02.112 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=496 oflag=direct 00:14:02.371 496+0 records in 00:14:02.371 496+0 records out 00:14:02.371 65011712 bytes (65 MB, 62 MiB) copied, 0.292684 s, 222 MB/s 00:14:02.371 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:02.371 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:02.371 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:02.371 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:02.371 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:02.371 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:02.371 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:02.631 [2024-12-05 19:04:19.981736] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:02.631 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:02.631 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:02.631 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:02.631 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:02.631 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:02.631 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:02.631 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:02.631 19:04:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.631 [2024-12-05 19:04:20.005777] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:02.631 "name": "raid_bdev1", 00:14:02.631 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:02.631 "strip_size_kb": 64, 00:14:02.631 "state": "online", 00:14:02.631 "raid_level": "raid5f", 00:14:02.631 "superblock": true, 00:14:02.631 "num_base_bdevs": 3, 00:14:02.631 "num_base_bdevs_discovered": 2, 00:14:02.631 "num_base_bdevs_operational": 2, 00:14:02.631 "base_bdevs_list": [ 00:14:02.631 { 00:14:02.631 "name": null, 00:14:02.631 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:02.631 "is_configured": false, 00:14:02.631 "data_offset": 0, 00:14:02.631 "data_size": 63488 00:14:02.631 }, 00:14:02.631 { 00:14:02.631 "name": "BaseBdev2", 00:14:02.631 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:02.631 "is_configured": true, 00:14:02.631 "data_offset": 2048, 00:14:02.631 "data_size": 63488 00:14:02.631 }, 00:14:02.631 { 00:14:02.631 "name": "BaseBdev3", 00:14:02.631 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:02.631 "is_configured": true, 00:14:02.631 "data_offset": 2048, 00:14:02.631 "data_size": 63488 00:14:02.631 } 00:14:02.631 ] 00:14:02.631 }' 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:02.631 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.891 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:02.891 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.891 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.891 [2024-12-05 19:04:20.445013] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:03.151 [2024-12-05 19:04:20.449721] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000255d0 00:14:03.151 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:03.151 19:04:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:03.151 [2024-12-05 19:04:20.451868] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:04.092 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:04.092 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:04.092 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:04.092 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:04.093 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:04.093 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.093 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:04.093 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.093 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.093 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.093 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:04.093 "name": "raid_bdev1", 00:14:04.093 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:04.093 "strip_size_kb": 64, 00:14:04.093 "state": "online", 00:14:04.093 "raid_level": "raid5f", 00:14:04.093 "superblock": true, 00:14:04.093 "num_base_bdevs": 3, 00:14:04.093 "num_base_bdevs_discovered": 3, 00:14:04.093 "num_base_bdevs_operational": 3, 00:14:04.093 "process": { 00:14:04.093 "type": "rebuild", 00:14:04.093 "target": "spare", 00:14:04.093 "progress": { 00:14:04.093 "blocks": 20480, 00:14:04.093 "percent": 16 00:14:04.093 } 00:14:04.093 }, 00:14:04.093 "base_bdevs_list": [ 00:14:04.093 { 00:14:04.093 "name": "spare", 00:14:04.093 "uuid": "58dbeef8-57fe-5922-abe0-dd0bddc541e0", 00:14:04.093 "is_configured": true, 00:14:04.093 "data_offset": 2048, 00:14:04.093 "data_size": 63488 00:14:04.093 }, 00:14:04.093 { 00:14:04.093 "name": "BaseBdev2", 00:14:04.093 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:04.093 "is_configured": true, 00:14:04.093 "data_offset": 2048, 00:14:04.093 "data_size": 63488 00:14:04.093 }, 00:14:04.093 { 00:14:04.093 "name": "BaseBdev3", 00:14:04.093 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:04.093 "is_configured": true, 00:14:04.093 "data_offset": 2048, 00:14:04.093 "data_size": 63488 00:14:04.093 } 00:14:04.093 ] 00:14:04.093 }' 00:14:04.093 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:04.093 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:04.093 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:04.093 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:04.093 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:04.093 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.093 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.093 [2024-12-05 19:04:21.607778] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:04.352 [2024-12-05 19:04:21.658560] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:04.352 [2024-12-05 19:04:21.658626] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:04.352 [2024-12-05 19:04:21.658641] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:04.352 [2024-12-05 19:04:21.658651] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.352 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:04.352 "name": "raid_bdev1", 00:14:04.352 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:04.352 "strip_size_kb": 64, 00:14:04.352 "state": "online", 00:14:04.352 "raid_level": "raid5f", 00:14:04.352 "superblock": true, 00:14:04.352 "num_base_bdevs": 3, 00:14:04.352 "num_base_bdevs_discovered": 2, 00:14:04.352 "num_base_bdevs_operational": 2, 00:14:04.352 "base_bdevs_list": [ 00:14:04.352 { 00:14:04.352 "name": null, 00:14:04.352 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:04.352 "is_configured": false, 00:14:04.352 "data_offset": 0, 00:14:04.352 "data_size": 63488 00:14:04.352 }, 00:14:04.352 { 00:14:04.352 "name": "BaseBdev2", 00:14:04.352 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:04.352 "is_configured": true, 00:14:04.352 "data_offset": 2048, 00:14:04.353 "data_size": 63488 00:14:04.353 }, 00:14:04.353 { 00:14:04.353 "name": "BaseBdev3", 00:14:04.353 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:04.353 "is_configured": true, 00:14:04.353 "data_offset": 2048, 00:14:04.353 "data_size": 63488 00:14:04.353 } 00:14:04.353 ] 00:14:04.353 }' 00:14:04.353 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:04.353 19:04:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.612 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:04.612 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:04.612 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:04.612 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:04.612 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:04.612 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:04.612 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.612 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.612 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.612 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.873 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:04.873 "name": "raid_bdev1", 00:14:04.873 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:04.873 "strip_size_kb": 64, 00:14:04.873 "state": "online", 00:14:04.873 "raid_level": "raid5f", 00:14:04.873 "superblock": true, 00:14:04.873 "num_base_bdevs": 3, 00:14:04.873 "num_base_bdevs_discovered": 2, 00:14:04.873 "num_base_bdevs_operational": 2, 00:14:04.873 "base_bdevs_list": [ 00:14:04.873 { 00:14:04.873 "name": null, 00:14:04.873 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:04.873 "is_configured": false, 00:14:04.873 "data_offset": 0, 00:14:04.873 "data_size": 63488 00:14:04.873 }, 00:14:04.873 { 00:14:04.873 "name": "BaseBdev2", 00:14:04.873 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:04.873 "is_configured": true, 00:14:04.873 "data_offset": 2048, 00:14:04.873 "data_size": 63488 00:14:04.873 }, 00:14:04.873 { 00:14:04.873 "name": "BaseBdev3", 00:14:04.873 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:04.873 "is_configured": true, 00:14:04.873 "data_offset": 2048, 00:14:04.873 "data_size": 63488 00:14:04.873 } 00:14:04.873 ] 00:14:04.873 }' 00:14:04.873 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:04.873 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:04.873 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:04.873 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:04.873 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:04.873 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:04.873 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.873 [2024-12-05 19:04:22.255720] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:04.873 [2024-12-05 19:04:22.259582] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000256a0 00:14:04.873 [2024-12-05 19:04:22.261723] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:04.873 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:04.873 19:04:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:05.813 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:05.813 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:05.813 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:05.813 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:05.814 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:05.814 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.814 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:05.814 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:05.814 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.814 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:05.814 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:05.814 "name": "raid_bdev1", 00:14:05.814 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:05.814 "strip_size_kb": 64, 00:14:05.814 "state": "online", 00:14:05.814 "raid_level": "raid5f", 00:14:05.814 "superblock": true, 00:14:05.814 "num_base_bdevs": 3, 00:14:05.814 "num_base_bdevs_discovered": 3, 00:14:05.814 "num_base_bdevs_operational": 3, 00:14:05.814 "process": { 00:14:05.814 "type": "rebuild", 00:14:05.814 "target": "spare", 00:14:05.814 "progress": { 00:14:05.814 "blocks": 20480, 00:14:05.814 "percent": 16 00:14:05.814 } 00:14:05.814 }, 00:14:05.814 "base_bdevs_list": [ 00:14:05.814 { 00:14:05.814 "name": "spare", 00:14:05.814 "uuid": "58dbeef8-57fe-5922-abe0-dd0bddc541e0", 00:14:05.814 "is_configured": true, 00:14:05.814 "data_offset": 2048, 00:14:05.814 "data_size": 63488 00:14:05.814 }, 00:14:05.814 { 00:14:05.814 "name": "BaseBdev2", 00:14:05.814 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:05.814 "is_configured": true, 00:14:05.814 "data_offset": 2048, 00:14:05.814 "data_size": 63488 00:14:05.814 }, 00:14:05.814 { 00:14:05.814 "name": "BaseBdev3", 00:14:05.814 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:05.814 "is_configured": true, 00:14:05.814 "data_offset": 2048, 00:14:05.814 "data_size": 63488 00:14:05.814 } 00:14:05.814 ] 00:14:05.814 }' 00:14:05.814 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:14:06.074 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=453 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:06.074 "name": "raid_bdev1", 00:14:06.074 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:06.074 "strip_size_kb": 64, 00:14:06.074 "state": "online", 00:14:06.074 "raid_level": "raid5f", 00:14:06.074 "superblock": true, 00:14:06.074 "num_base_bdevs": 3, 00:14:06.074 "num_base_bdevs_discovered": 3, 00:14:06.074 "num_base_bdevs_operational": 3, 00:14:06.074 "process": { 00:14:06.074 "type": "rebuild", 00:14:06.074 "target": "spare", 00:14:06.074 "progress": { 00:14:06.074 "blocks": 22528, 00:14:06.074 "percent": 17 00:14:06.074 } 00:14:06.074 }, 00:14:06.074 "base_bdevs_list": [ 00:14:06.074 { 00:14:06.074 "name": "spare", 00:14:06.074 "uuid": "58dbeef8-57fe-5922-abe0-dd0bddc541e0", 00:14:06.074 "is_configured": true, 00:14:06.074 "data_offset": 2048, 00:14:06.074 "data_size": 63488 00:14:06.074 }, 00:14:06.074 { 00:14:06.074 "name": "BaseBdev2", 00:14:06.074 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:06.074 "is_configured": true, 00:14:06.074 "data_offset": 2048, 00:14:06.074 "data_size": 63488 00:14:06.074 }, 00:14:06.074 { 00:14:06.074 "name": "BaseBdev3", 00:14:06.074 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:06.074 "is_configured": true, 00:14:06.074 "data_offset": 2048, 00:14:06.074 "data_size": 63488 00:14:06.074 } 00:14:06.074 ] 00:14:06.074 }' 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:06.074 19:04:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:07.457 "name": "raid_bdev1", 00:14:07.457 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:07.457 "strip_size_kb": 64, 00:14:07.457 "state": "online", 00:14:07.457 "raid_level": "raid5f", 00:14:07.457 "superblock": true, 00:14:07.457 "num_base_bdevs": 3, 00:14:07.457 "num_base_bdevs_discovered": 3, 00:14:07.457 "num_base_bdevs_operational": 3, 00:14:07.457 "process": { 00:14:07.457 "type": "rebuild", 00:14:07.457 "target": "spare", 00:14:07.457 "progress": { 00:14:07.457 "blocks": 47104, 00:14:07.457 "percent": 37 00:14:07.457 } 00:14:07.457 }, 00:14:07.457 "base_bdevs_list": [ 00:14:07.457 { 00:14:07.457 "name": "spare", 00:14:07.457 "uuid": "58dbeef8-57fe-5922-abe0-dd0bddc541e0", 00:14:07.457 "is_configured": true, 00:14:07.457 "data_offset": 2048, 00:14:07.457 "data_size": 63488 00:14:07.457 }, 00:14:07.457 { 00:14:07.457 "name": "BaseBdev2", 00:14:07.457 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:07.457 "is_configured": true, 00:14:07.457 "data_offset": 2048, 00:14:07.457 "data_size": 63488 00:14:07.457 }, 00:14:07.457 { 00:14:07.457 "name": "BaseBdev3", 00:14:07.457 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:07.457 "is_configured": true, 00:14:07.457 "data_offset": 2048, 00:14:07.457 "data_size": 63488 00:14:07.457 } 00:14:07.457 ] 00:14:07.457 }' 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:07.457 19:04:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:08.394 "name": "raid_bdev1", 00:14:08.394 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:08.394 "strip_size_kb": 64, 00:14:08.394 "state": "online", 00:14:08.394 "raid_level": "raid5f", 00:14:08.394 "superblock": true, 00:14:08.394 "num_base_bdevs": 3, 00:14:08.394 "num_base_bdevs_discovered": 3, 00:14:08.394 "num_base_bdevs_operational": 3, 00:14:08.394 "process": { 00:14:08.394 "type": "rebuild", 00:14:08.394 "target": "spare", 00:14:08.394 "progress": { 00:14:08.394 "blocks": 69632, 00:14:08.394 "percent": 54 00:14:08.394 } 00:14:08.394 }, 00:14:08.394 "base_bdevs_list": [ 00:14:08.394 { 00:14:08.394 "name": "spare", 00:14:08.394 "uuid": "58dbeef8-57fe-5922-abe0-dd0bddc541e0", 00:14:08.394 "is_configured": true, 00:14:08.394 "data_offset": 2048, 00:14:08.394 "data_size": 63488 00:14:08.394 }, 00:14:08.394 { 00:14:08.394 "name": "BaseBdev2", 00:14:08.394 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:08.394 "is_configured": true, 00:14:08.394 "data_offset": 2048, 00:14:08.394 "data_size": 63488 00:14:08.394 }, 00:14:08.394 { 00:14:08.394 "name": "BaseBdev3", 00:14:08.394 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:08.394 "is_configured": true, 00:14:08.394 "data_offset": 2048, 00:14:08.394 "data_size": 63488 00:14:08.394 } 00:14:08.394 ] 00:14:08.394 }' 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:08.394 19:04:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:09.342 19:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:09.342 19:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:09.342 19:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:09.342 19:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:09.342 19:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:09.342 19:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:09.342 19:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.342 19:04:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.342 19:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:09.342 19:04:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:09.342 19:04:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.601 19:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:09.601 "name": "raid_bdev1", 00:14:09.601 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:09.601 "strip_size_kb": 64, 00:14:09.601 "state": "online", 00:14:09.601 "raid_level": "raid5f", 00:14:09.601 "superblock": true, 00:14:09.601 "num_base_bdevs": 3, 00:14:09.601 "num_base_bdevs_discovered": 3, 00:14:09.601 "num_base_bdevs_operational": 3, 00:14:09.601 "process": { 00:14:09.601 "type": "rebuild", 00:14:09.601 "target": "spare", 00:14:09.601 "progress": { 00:14:09.601 "blocks": 92160, 00:14:09.601 "percent": 72 00:14:09.601 } 00:14:09.601 }, 00:14:09.601 "base_bdevs_list": [ 00:14:09.601 { 00:14:09.601 "name": "spare", 00:14:09.601 "uuid": "58dbeef8-57fe-5922-abe0-dd0bddc541e0", 00:14:09.601 "is_configured": true, 00:14:09.601 "data_offset": 2048, 00:14:09.601 "data_size": 63488 00:14:09.601 }, 00:14:09.601 { 00:14:09.601 "name": "BaseBdev2", 00:14:09.601 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:09.601 "is_configured": true, 00:14:09.601 "data_offset": 2048, 00:14:09.601 "data_size": 63488 00:14:09.601 }, 00:14:09.601 { 00:14:09.601 "name": "BaseBdev3", 00:14:09.601 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:09.601 "is_configured": true, 00:14:09.601 "data_offset": 2048, 00:14:09.601 "data_size": 63488 00:14:09.601 } 00:14:09.601 ] 00:14:09.601 }' 00:14:09.601 19:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:09.601 19:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:09.601 19:04:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:09.601 19:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:09.601 19:04:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:10.538 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:10.538 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:10.538 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:10.538 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:10.538 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:10.538 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:10.538 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:10.538 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:10.538 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:10.538 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:10.538 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:10.538 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:10.538 "name": "raid_bdev1", 00:14:10.538 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:10.538 "strip_size_kb": 64, 00:14:10.538 "state": "online", 00:14:10.538 "raid_level": "raid5f", 00:14:10.538 "superblock": true, 00:14:10.538 "num_base_bdevs": 3, 00:14:10.538 "num_base_bdevs_discovered": 3, 00:14:10.538 "num_base_bdevs_operational": 3, 00:14:10.538 "process": { 00:14:10.538 "type": "rebuild", 00:14:10.538 "target": "spare", 00:14:10.538 "progress": { 00:14:10.538 "blocks": 116736, 00:14:10.538 "percent": 91 00:14:10.538 } 00:14:10.538 }, 00:14:10.538 "base_bdevs_list": [ 00:14:10.538 { 00:14:10.538 "name": "spare", 00:14:10.538 "uuid": "58dbeef8-57fe-5922-abe0-dd0bddc541e0", 00:14:10.538 "is_configured": true, 00:14:10.538 "data_offset": 2048, 00:14:10.538 "data_size": 63488 00:14:10.538 }, 00:14:10.538 { 00:14:10.538 "name": "BaseBdev2", 00:14:10.538 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:10.538 "is_configured": true, 00:14:10.538 "data_offset": 2048, 00:14:10.538 "data_size": 63488 00:14:10.538 }, 00:14:10.538 { 00:14:10.538 "name": "BaseBdev3", 00:14:10.538 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:10.538 "is_configured": true, 00:14:10.538 "data_offset": 2048, 00:14:10.538 "data_size": 63488 00:14:10.538 } 00:14:10.538 ] 00:14:10.538 }' 00:14:10.538 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:10.797 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:10.797 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:10.797 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:10.797 19:04:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:11.056 [2024-12-05 19:04:28.493835] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:11.056 [2024-12-05 19:04:28.493900] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:11.056 [2024-12-05 19:04:28.493999] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:11.626 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:11.626 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:11.626 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:11.626 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:11.626 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:11.626 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:11.626 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.626 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.626 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:11.626 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.626 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:11.885 "name": "raid_bdev1", 00:14:11.885 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:11.885 "strip_size_kb": 64, 00:14:11.885 "state": "online", 00:14:11.885 "raid_level": "raid5f", 00:14:11.885 "superblock": true, 00:14:11.885 "num_base_bdevs": 3, 00:14:11.885 "num_base_bdevs_discovered": 3, 00:14:11.885 "num_base_bdevs_operational": 3, 00:14:11.885 "base_bdevs_list": [ 00:14:11.885 { 00:14:11.885 "name": "spare", 00:14:11.885 "uuid": "58dbeef8-57fe-5922-abe0-dd0bddc541e0", 00:14:11.885 "is_configured": true, 00:14:11.885 "data_offset": 2048, 00:14:11.885 "data_size": 63488 00:14:11.885 }, 00:14:11.885 { 00:14:11.885 "name": "BaseBdev2", 00:14:11.885 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:11.885 "is_configured": true, 00:14:11.885 "data_offset": 2048, 00:14:11.885 "data_size": 63488 00:14:11.885 }, 00:14:11.885 { 00:14:11.885 "name": "BaseBdev3", 00:14:11.885 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:11.885 "is_configured": true, 00:14:11.885 "data_offset": 2048, 00:14:11.885 "data_size": 63488 00:14:11.885 } 00:14:11.885 ] 00:14:11.885 }' 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:11.885 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:11.885 "name": "raid_bdev1", 00:14:11.885 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:11.885 "strip_size_kb": 64, 00:14:11.885 "state": "online", 00:14:11.885 "raid_level": "raid5f", 00:14:11.885 "superblock": true, 00:14:11.885 "num_base_bdevs": 3, 00:14:11.885 "num_base_bdevs_discovered": 3, 00:14:11.885 "num_base_bdevs_operational": 3, 00:14:11.885 "base_bdevs_list": [ 00:14:11.885 { 00:14:11.885 "name": "spare", 00:14:11.885 "uuid": "58dbeef8-57fe-5922-abe0-dd0bddc541e0", 00:14:11.885 "is_configured": true, 00:14:11.885 "data_offset": 2048, 00:14:11.885 "data_size": 63488 00:14:11.885 }, 00:14:11.885 { 00:14:11.885 "name": "BaseBdev2", 00:14:11.885 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:11.885 "is_configured": true, 00:14:11.885 "data_offset": 2048, 00:14:11.885 "data_size": 63488 00:14:11.885 }, 00:14:11.885 { 00:14:11.885 "name": "BaseBdev3", 00:14:11.885 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:11.885 "is_configured": true, 00:14:11.885 "data_offset": 2048, 00:14:11.885 "data_size": 63488 00:14:11.885 } 00:14:11.885 ] 00:14:11.886 }' 00:14:11.886 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:11.886 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:11.886 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:12.144 "name": "raid_bdev1", 00:14:12.144 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:12.144 "strip_size_kb": 64, 00:14:12.144 "state": "online", 00:14:12.144 "raid_level": "raid5f", 00:14:12.144 "superblock": true, 00:14:12.144 "num_base_bdevs": 3, 00:14:12.144 "num_base_bdevs_discovered": 3, 00:14:12.144 "num_base_bdevs_operational": 3, 00:14:12.144 "base_bdevs_list": [ 00:14:12.144 { 00:14:12.144 "name": "spare", 00:14:12.144 "uuid": "58dbeef8-57fe-5922-abe0-dd0bddc541e0", 00:14:12.144 "is_configured": true, 00:14:12.144 "data_offset": 2048, 00:14:12.144 "data_size": 63488 00:14:12.144 }, 00:14:12.144 { 00:14:12.144 "name": "BaseBdev2", 00:14:12.144 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:12.144 "is_configured": true, 00:14:12.144 "data_offset": 2048, 00:14:12.144 "data_size": 63488 00:14:12.144 }, 00:14:12.144 { 00:14:12.144 "name": "BaseBdev3", 00:14:12.144 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:12.144 "is_configured": true, 00:14:12.144 "data_offset": 2048, 00:14:12.144 "data_size": 63488 00:14:12.144 } 00:14:12.144 ] 00:14:12.144 }' 00:14:12.144 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:12.145 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.404 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:12.404 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.404 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.404 [2024-12-05 19:04:29.936871] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:12.404 [2024-12-05 19:04:29.936917] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:12.404 [2024-12-05 19:04:29.936982] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:12.404 [2024-12-05 19:04:29.937052] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:12.404 [2024-12-05 19:04:29.937093] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:12.404 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.404 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.404 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:12.404 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:14:12.404 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:12.404 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:12.685 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:12.685 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:12.685 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:12.685 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:12.685 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:12.685 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:12.685 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:12.685 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:12.685 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:12.685 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:12.685 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:12.685 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:12.685 19:04:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:12.685 /dev/nbd0 00:14:12.685 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:12.685 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:12.685 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:12.685 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:14:12.685 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:12.685 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:12.685 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:12.685 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:14:12.685 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:12.685 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:12.685 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:12.944 1+0 records in 00:14:12.944 1+0 records out 00:14:12.944 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000403627 s, 10.1 MB/s 00:14:12.944 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:12.944 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:14:12.944 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:12.944 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:12.944 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:14:12.944 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:12.945 /dev/nbd1 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:12.945 1+0 records in 00:14:12.945 1+0 records out 00:14:12.945 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000251692 s, 16.3 MB/s 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:12.945 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:14:13.204 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:13.204 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:13.204 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:13.204 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:13.204 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:13.204 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:13.204 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.463 [2024-12-05 19:04:30.980163] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:13.463 [2024-12-05 19:04:30.980225] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:13.463 [2024-12-05 19:04:30.980247] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:14:13.463 [2024-12-05 19:04:30.980258] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:13.463 [2024-12-05 19:04:30.982273] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:13.463 [2024-12-05 19:04:30.982311] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:13.463 [2024-12-05 19:04:30.982389] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:13.463 [2024-12-05 19:04:30.982428] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:13.463 [2024-12-05 19:04:30.982551] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:13.463 [2024-12-05 19:04:30.982664] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:13.463 spare 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.463 19:04:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.722 [2024-12-05 19:04:31.082574] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:14:13.722 [2024-12-05 19:04:31.082602] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:13.722 [2024-12-05 19:04:31.082904] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043d50 00:14:13.722 [2024-12-05 19:04:31.083327] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:14:13.722 [2024-12-05 19:04:31.083350] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:14:13.722 [2024-12-05 19:04:31.083482] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:13.722 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:13.722 "name": "raid_bdev1", 00:14:13.722 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:13.722 "strip_size_kb": 64, 00:14:13.722 "state": "online", 00:14:13.722 "raid_level": "raid5f", 00:14:13.722 "superblock": true, 00:14:13.722 "num_base_bdevs": 3, 00:14:13.722 "num_base_bdevs_discovered": 3, 00:14:13.722 "num_base_bdevs_operational": 3, 00:14:13.722 "base_bdevs_list": [ 00:14:13.722 { 00:14:13.722 "name": "spare", 00:14:13.722 "uuid": "58dbeef8-57fe-5922-abe0-dd0bddc541e0", 00:14:13.722 "is_configured": true, 00:14:13.722 "data_offset": 2048, 00:14:13.722 "data_size": 63488 00:14:13.722 }, 00:14:13.722 { 00:14:13.722 "name": "BaseBdev2", 00:14:13.722 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:13.722 "is_configured": true, 00:14:13.722 "data_offset": 2048, 00:14:13.722 "data_size": 63488 00:14:13.722 }, 00:14:13.722 { 00:14:13.722 "name": "BaseBdev3", 00:14:13.722 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:13.722 "is_configured": true, 00:14:13.722 "data_offset": 2048, 00:14:13.723 "data_size": 63488 00:14:13.723 } 00:14:13.723 ] 00:14:13.723 }' 00:14:13.723 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:13.723 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:14.291 "name": "raid_bdev1", 00:14:14.291 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:14.291 "strip_size_kb": 64, 00:14:14.291 "state": "online", 00:14:14.291 "raid_level": "raid5f", 00:14:14.291 "superblock": true, 00:14:14.291 "num_base_bdevs": 3, 00:14:14.291 "num_base_bdevs_discovered": 3, 00:14:14.291 "num_base_bdevs_operational": 3, 00:14:14.291 "base_bdevs_list": [ 00:14:14.291 { 00:14:14.291 "name": "spare", 00:14:14.291 "uuid": "58dbeef8-57fe-5922-abe0-dd0bddc541e0", 00:14:14.291 "is_configured": true, 00:14:14.291 "data_offset": 2048, 00:14:14.291 "data_size": 63488 00:14:14.291 }, 00:14:14.291 { 00:14:14.291 "name": "BaseBdev2", 00:14:14.291 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:14.291 "is_configured": true, 00:14:14.291 "data_offset": 2048, 00:14:14.291 "data_size": 63488 00:14:14.291 }, 00:14:14.291 { 00:14:14.291 "name": "BaseBdev3", 00:14:14.291 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:14.291 "is_configured": true, 00:14:14.291 "data_offset": 2048, 00:14:14.291 "data_size": 63488 00:14:14.291 } 00:14:14.291 ] 00:14:14.291 }' 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.291 [2024-12-05 19:04:31.759629] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:14.291 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.292 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.292 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.292 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:14.292 "name": "raid_bdev1", 00:14:14.292 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:14.292 "strip_size_kb": 64, 00:14:14.292 "state": "online", 00:14:14.292 "raid_level": "raid5f", 00:14:14.292 "superblock": true, 00:14:14.292 "num_base_bdevs": 3, 00:14:14.292 "num_base_bdevs_discovered": 2, 00:14:14.292 "num_base_bdevs_operational": 2, 00:14:14.292 "base_bdevs_list": [ 00:14:14.292 { 00:14:14.292 "name": null, 00:14:14.292 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:14.292 "is_configured": false, 00:14:14.292 "data_offset": 0, 00:14:14.292 "data_size": 63488 00:14:14.292 }, 00:14:14.292 { 00:14:14.292 "name": "BaseBdev2", 00:14:14.292 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:14.292 "is_configured": true, 00:14:14.292 "data_offset": 2048, 00:14:14.292 "data_size": 63488 00:14:14.292 }, 00:14:14.292 { 00:14:14.292 "name": "BaseBdev3", 00:14:14.292 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:14.292 "is_configured": true, 00:14:14.292 "data_offset": 2048, 00:14:14.292 "data_size": 63488 00:14:14.292 } 00:14:14.292 ] 00:14:14.292 }' 00:14:14.292 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:14.292 19:04:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.861 19:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:14.861 19:04:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:14.861 19:04:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:14.861 [2024-12-05 19:04:32.242813] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:14.861 [2024-12-05 19:04:32.242972] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:14:14.861 [2024-12-05 19:04:32.242996] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:14.861 [2024-12-05 19:04:32.243029] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:14.861 [2024-12-05 19:04:32.247573] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043e20 00:14:14.861 19:04:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:14.861 19:04:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:14:14.861 [2024-12-05 19:04:32.249698] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:15.801 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:15.801 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:15.801 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:15.801 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:15.801 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:15.801 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.801 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:15.801 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:15.801 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:15.801 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:15.801 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:15.801 "name": "raid_bdev1", 00:14:15.801 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:15.801 "strip_size_kb": 64, 00:14:15.801 "state": "online", 00:14:15.801 "raid_level": "raid5f", 00:14:15.801 "superblock": true, 00:14:15.801 "num_base_bdevs": 3, 00:14:15.801 "num_base_bdevs_discovered": 3, 00:14:15.801 "num_base_bdevs_operational": 3, 00:14:15.801 "process": { 00:14:15.801 "type": "rebuild", 00:14:15.801 "target": "spare", 00:14:15.801 "progress": { 00:14:15.801 "blocks": 20480, 00:14:15.801 "percent": 16 00:14:15.801 } 00:14:15.801 }, 00:14:15.801 "base_bdevs_list": [ 00:14:15.801 { 00:14:15.801 "name": "spare", 00:14:15.802 "uuid": "58dbeef8-57fe-5922-abe0-dd0bddc541e0", 00:14:15.802 "is_configured": true, 00:14:15.802 "data_offset": 2048, 00:14:15.802 "data_size": 63488 00:14:15.802 }, 00:14:15.802 { 00:14:15.802 "name": "BaseBdev2", 00:14:15.802 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:15.802 "is_configured": true, 00:14:15.802 "data_offset": 2048, 00:14:15.802 "data_size": 63488 00:14:15.802 }, 00:14:15.802 { 00:14:15.802 "name": "BaseBdev3", 00:14:15.802 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:15.802 "is_configured": true, 00:14:15.802 "data_offset": 2048, 00:14:15.802 "data_size": 63488 00:14:15.802 } 00:14:15.802 ] 00:14:15.802 }' 00:14:15.802 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:15.802 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:15.802 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:16.063 [2024-12-05 19:04:33.389535] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:16.063 [2024-12-05 19:04:33.456238] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:16.063 [2024-12-05 19:04:33.456294] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:16.063 [2024-12-05 19:04:33.456328] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:16.063 [2024-12-05 19:04:33.456335] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:16.063 "name": "raid_bdev1", 00:14:16.063 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:16.063 "strip_size_kb": 64, 00:14:16.063 "state": "online", 00:14:16.063 "raid_level": "raid5f", 00:14:16.063 "superblock": true, 00:14:16.063 "num_base_bdevs": 3, 00:14:16.063 "num_base_bdevs_discovered": 2, 00:14:16.063 "num_base_bdevs_operational": 2, 00:14:16.063 "base_bdevs_list": [ 00:14:16.063 { 00:14:16.063 "name": null, 00:14:16.063 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:16.063 "is_configured": false, 00:14:16.063 "data_offset": 0, 00:14:16.063 "data_size": 63488 00:14:16.063 }, 00:14:16.063 { 00:14:16.063 "name": "BaseBdev2", 00:14:16.063 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:16.063 "is_configured": true, 00:14:16.063 "data_offset": 2048, 00:14:16.063 "data_size": 63488 00:14:16.063 }, 00:14:16.063 { 00:14:16.063 "name": "BaseBdev3", 00:14:16.063 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:16.063 "is_configured": true, 00:14:16.063 "data_offset": 2048, 00:14:16.063 "data_size": 63488 00:14:16.063 } 00:14:16.063 ] 00:14:16.063 }' 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:16.063 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:16.632 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:16.633 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:16.633 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:16.633 [2024-12-05 19:04:33.932837] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:16.633 [2024-12-05 19:04:33.932891] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:16.633 [2024-12-05 19:04:33.932912] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:14:16.633 [2024-12-05 19:04:33.932921] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:16.633 [2024-12-05 19:04:33.933353] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:16.633 [2024-12-05 19:04:33.933385] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:16.633 [2024-12-05 19:04:33.933458] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:16.633 [2024-12-05 19:04:33.933480] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:14:16.633 [2024-12-05 19:04:33.933494] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:16.633 [2024-12-05 19:04:33.933528] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:16.633 [2024-12-05 19:04:33.936963] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043ef0 00:14:16.633 spare 00:14:16.633 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:16.633 19:04:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:14:16.633 [2024-12-05 19:04:33.939041] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:17.574 19:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:17.574 19:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:17.574 19:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:17.574 19:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:17.574 19:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:17.574 19:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:17.574 19:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:17.574 19:04:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.574 19:04:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.574 19:04:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.574 19:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:17.574 "name": "raid_bdev1", 00:14:17.574 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:17.574 "strip_size_kb": 64, 00:14:17.574 "state": "online", 00:14:17.574 "raid_level": "raid5f", 00:14:17.574 "superblock": true, 00:14:17.574 "num_base_bdevs": 3, 00:14:17.574 "num_base_bdevs_discovered": 3, 00:14:17.574 "num_base_bdevs_operational": 3, 00:14:17.574 "process": { 00:14:17.574 "type": "rebuild", 00:14:17.574 "target": "spare", 00:14:17.574 "progress": { 00:14:17.574 "blocks": 20480, 00:14:17.574 "percent": 16 00:14:17.574 } 00:14:17.574 }, 00:14:17.574 "base_bdevs_list": [ 00:14:17.574 { 00:14:17.574 "name": "spare", 00:14:17.574 "uuid": "58dbeef8-57fe-5922-abe0-dd0bddc541e0", 00:14:17.574 "is_configured": true, 00:14:17.574 "data_offset": 2048, 00:14:17.574 "data_size": 63488 00:14:17.574 }, 00:14:17.574 { 00:14:17.574 "name": "BaseBdev2", 00:14:17.574 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:17.574 "is_configured": true, 00:14:17.574 "data_offset": 2048, 00:14:17.574 "data_size": 63488 00:14:17.574 }, 00:14:17.574 { 00:14:17.574 "name": "BaseBdev3", 00:14:17.575 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:17.575 "is_configured": true, 00:14:17.575 "data_offset": 2048, 00:14:17.575 "data_size": 63488 00:14:17.575 } 00:14:17.575 ] 00:14:17.575 }' 00:14:17.575 19:04:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:17.575 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:17.575 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:17.575 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:17.575 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:14:17.575 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.575 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.575 [2024-12-05 19:04:35.083625] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:17.834 [2024-12-05 19:04:35.145433] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:17.834 [2024-12-05 19:04:35.145489] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:17.834 [2024-12-05 19:04:35.145520] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:17.834 [2024-12-05 19:04:35.145531] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:17.834 "name": "raid_bdev1", 00:14:17.834 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:17.834 "strip_size_kb": 64, 00:14:17.834 "state": "online", 00:14:17.834 "raid_level": "raid5f", 00:14:17.834 "superblock": true, 00:14:17.834 "num_base_bdevs": 3, 00:14:17.834 "num_base_bdevs_discovered": 2, 00:14:17.834 "num_base_bdevs_operational": 2, 00:14:17.834 "base_bdevs_list": [ 00:14:17.834 { 00:14:17.834 "name": null, 00:14:17.834 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:17.834 "is_configured": false, 00:14:17.834 "data_offset": 0, 00:14:17.834 "data_size": 63488 00:14:17.834 }, 00:14:17.834 { 00:14:17.834 "name": "BaseBdev2", 00:14:17.834 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:17.834 "is_configured": true, 00:14:17.834 "data_offset": 2048, 00:14:17.834 "data_size": 63488 00:14:17.834 }, 00:14:17.834 { 00:14:17.834 "name": "BaseBdev3", 00:14:17.834 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:17.834 "is_configured": true, 00:14:17.834 "data_offset": 2048, 00:14:17.834 "data_size": 63488 00:14:17.834 } 00:14:17.834 ] 00:14:17.834 }' 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:17.834 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.093 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:18.093 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:18.093 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:18.093 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:18.093 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:18.093 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:18.093 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.093 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.093 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:18.093 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.352 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:18.352 "name": "raid_bdev1", 00:14:18.352 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:18.352 "strip_size_kb": 64, 00:14:18.352 "state": "online", 00:14:18.352 "raid_level": "raid5f", 00:14:18.352 "superblock": true, 00:14:18.352 "num_base_bdevs": 3, 00:14:18.352 "num_base_bdevs_discovered": 2, 00:14:18.352 "num_base_bdevs_operational": 2, 00:14:18.352 "base_bdevs_list": [ 00:14:18.352 { 00:14:18.352 "name": null, 00:14:18.352 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:18.352 "is_configured": false, 00:14:18.352 "data_offset": 0, 00:14:18.352 "data_size": 63488 00:14:18.352 }, 00:14:18.352 { 00:14:18.352 "name": "BaseBdev2", 00:14:18.352 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:18.352 "is_configured": true, 00:14:18.352 "data_offset": 2048, 00:14:18.352 "data_size": 63488 00:14:18.352 }, 00:14:18.352 { 00:14:18.352 "name": "BaseBdev3", 00:14:18.352 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:18.352 "is_configured": true, 00:14:18.352 "data_offset": 2048, 00:14:18.352 "data_size": 63488 00:14:18.352 } 00:14:18.352 ] 00:14:18.352 }' 00:14:18.352 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:18.352 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:18.352 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:18.352 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:18.352 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:14:18.352 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.352 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.352 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.352 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:18.352 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:18.352 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.352 [2024-12-05 19:04:35.781723] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:18.352 [2024-12-05 19:04:35.781776] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:18.352 [2024-12-05 19:04:35.781795] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:14:18.352 [2024-12-05 19:04:35.781806] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:18.352 [2024-12-05 19:04:35.782206] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:18.352 [2024-12-05 19:04:35.782243] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:18.352 [2024-12-05 19:04:35.782303] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:14:18.352 [2024-12-05 19:04:35.782320] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:14:18.352 [2024-12-05 19:04:35.782329] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:18.352 [2024-12-05 19:04:35.782341] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:14:18.352 BaseBdev1 00:14:18.352 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:18.352 19:04:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:14:19.292 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:19.292 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:19.292 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:19.292 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:19.292 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:19.292 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:19.293 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:19.293 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:19.293 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:19.293 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:19.293 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:19.293 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.293 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:19.293 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.293 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.293 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:19.293 "name": "raid_bdev1", 00:14:19.293 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:19.293 "strip_size_kb": 64, 00:14:19.293 "state": "online", 00:14:19.293 "raid_level": "raid5f", 00:14:19.293 "superblock": true, 00:14:19.293 "num_base_bdevs": 3, 00:14:19.293 "num_base_bdevs_discovered": 2, 00:14:19.293 "num_base_bdevs_operational": 2, 00:14:19.293 "base_bdevs_list": [ 00:14:19.293 { 00:14:19.293 "name": null, 00:14:19.293 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:19.293 "is_configured": false, 00:14:19.293 "data_offset": 0, 00:14:19.293 "data_size": 63488 00:14:19.293 }, 00:14:19.293 { 00:14:19.293 "name": "BaseBdev2", 00:14:19.293 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:19.293 "is_configured": true, 00:14:19.293 "data_offset": 2048, 00:14:19.293 "data_size": 63488 00:14:19.293 }, 00:14:19.293 { 00:14:19.293 "name": "BaseBdev3", 00:14:19.293 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:19.293 "is_configured": true, 00:14:19.293 "data_offset": 2048, 00:14:19.293 "data_size": 63488 00:14:19.293 } 00:14:19.293 ] 00:14:19.293 }' 00:14:19.293 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:19.293 19:04:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:19.861 "name": "raid_bdev1", 00:14:19.861 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:19.861 "strip_size_kb": 64, 00:14:19.861 "state": "online", 00:14:19.861 "raid_level": "raid5f", 00:14:19.861 "superblock": true, 00:14:19.861 "num_base_bdevs": 3, 00:14:19.861 "num_base_bdevs_discovered": 2, 00:14:19.861 "num_base_bdevs_operational": 2, 00:14:19.861 "base_bdevs_list": [ 00:14:19.861 { 00:14:19.861 "name": null, 00:14:19.861 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:19.861 "is_configured": false, 00:14:19.861 "data_offset": 0, 00:14:19.861 "data_size": 63488 00:14:19.861 }, 00:14:19.861 { 00:14:19.861 "name": "BaseBdev2", 00:14:19.861 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:19.861 "is_configured": true, 00:14:19.861 "data_offset": 2048, 00:14:19.861 "data_size": 63488 00:14:19.861 }, 00:14:19.861 { 00:14:19.861 "name": "BaseBdev3", 00:14:19.861 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:19.861 "is_configured": true, 00:14:19.861 "data_offset": 2048, 00:14:19.861 "data_size": 63488 00:14:19.861 } 00:14:19.861 ] 00:14:19.861 }' 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:19.861 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.862 [2024-12-05 19:04:37.379032] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:19.862 [2024-12-05 19:04:37.379153] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:14:19.862 [2024-12-05 19:04:37.379169] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:19.862 request: 00:14:19.862 { 00:14:19.862 "base_bdev": "BaseBdev1", 00:14:19.862 "raid_bdev": "raid_bdev1", 00:14:19.862 "method": "bdev_raid_add_base_bdev", 00:14:19.862 "req_id": 1 00:14:19.862 } 00:14:19.862 Got JSON-RPC error response 00:14:19.862 response: 00:14:19.862 { 00:14:19.862 "code": -22, 00:14:19.862 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:14:19.862 } 00:14:19.862 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:14:19.862 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:14:19.862 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:19.862 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:19.862 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:19.862 19:04:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:21.243 "name": "raid_bdev1", 00:14:21.243 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:21.243 "strip_size_kb": 64, 00:14:21.243 "state": "online", 00:14:21.243 "raid_level": "raid5f", 00:14:21.243 "superblock": true, 00:14:21.243 "num_base_bdevs": 3, 00:14:21.243 "num_base_bdevs_discovered": 2, 00:14:21.243 "num_base_bdevs_operational": 2, 00:14:21.243 "base_bdevs_list": [ 00:14:21.243 { 00:14:21.243 "name": null, 00:14:21.243 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:21.243 "is_configured": false, 00:14:21.243 "data_offset": 0, 00:14:21.243 "data_size": 63488 00:14:21.243 }, 00:14:21.243 { 00:14:21.243 "name": "BaseBdev2", 00:14:21.243 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:21.243 "is_configured": true, 00:14:21.243 "data_offset": 2048, 00:14:21.243 "data_size": 63488 00:14:21.243 }, 00:14:21.243 { 00:14:21.243 "name": "BaseBdev3", 00:14:21.243 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:21.243 "is_configured": true, 00:14:21.243 "data_offset": 2048, 00:14:21.243 "data_size": 63488 00:14:21.243 } 00:14:21.243 ] 00:14:21.243 }' 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:21.243 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.502 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:21.502 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:21.502 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:21.502 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:21.502 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:21.502 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:21.502 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:21.502 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:21.502 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.502 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:21.502 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:21.502 "name": "raid_bdev1", 00:14:21.502 "uuid": "573701a6-5809-4061-a5fc-96ec86d1432e", 00:14:21.502 "strip_size_kb": 64, 00:14:21.502 "state": "online", 00:14:21.502 "raid_level": "raid5f", 00:14:21.502 "superblock": true, 00:14:21.502 "num_base_bdevs": 3, 00:14:21.502 "num_base_bdevs_discovered": 2, 00:14:21.502 "num_base_bdevs_operational": 2, 00:14:21.502 "base_bdevs_list": [ 00:14:21.502 { 00:14:21.502 "name": null, 00:14:21.502 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:21.502 "is_configured": false, 00:14:21.502 "data_offset": 0, 00:14:21.502 "data_size": 63488 00:14:21.502 }, 00:14:21.502 { 00:14:21.502 "name": "BaseBdev2", 00:14:21.502 "uuid": "e4fdfd3b-23ef-5453-9116-dfb4a231258e", 00:14:21.502 "is_configured": true, 00:14:21.502 "data_offset": 2048, 00:14:21.502 "data_size": 63488 00:14:21.502 }, 00:14:21.502 { 00:14:21.502 "name": "BaseBdev3", 00:14:21.502 "uuid": "c47c93ee-4231-5210-b806-d78b988b531d", 00:14:21.502 "is_configured": true, 00:14:21.502 "data_offset": 2048, 00:14:21.502 "data_size": 63488 00:14:21.502 } 00:14:21.502 ] 00:14:21.502 }' 00:14:21.503 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:21.503 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:21.503 19:04:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:21.503 19:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:21.503 19:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 90244 00:14:21.503 19:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 90244 ']' 00:14:21.503 19:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 90244 00:14:21.503 19:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:14:21.503 19:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:21.503 19:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 90244 00:14:21.503 killing process with pid 90244 00:14:21.503 Received shutdown signal, test time was about 60.000000 seconds 00:14:21.503 00:14:21.503 Latency(us) 00:14:21.503 [2024-12-05T19:04:39.062Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:21.503 [2024-12-05T19:04:39.062Z] =================================================================================================================== 00:14:21.503 [2024-12-05T19:04:39.062Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:21.503 19:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:21.503 19:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:21.503 19:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 90244' 00:14:21.503 19:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 90244 00:14:21.503 [2024-12-05 19:04:39.043963] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:21.503 [2024-12-05 19:04:39.044083] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:21.503 19:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 90244 00:14:21.503 [2024-12-05 19:04:39.044150] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:21.503 [2024-12-05 19:04:39.044162] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:14:21.762 [2024-12-05 19:04:39.084652] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:21.762 19:04:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:14:21.762 00:14:21.762 real 0m21.646s 00:14:21.762 user 0m28.325s 00:14:21.762 sys 0m2.699s 00:14:21.762 19:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:21.762 19:04:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.762 ************************************ 00:14:21.762 END TEST raid5f_rebuild_test_sb 00:14:21.762 ************************************ 00:14:22.022 19:04:39 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:14:22.022 19:04:39 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 4 false 00:14:22.022 19:04:39 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:14:22.022 19:04:39 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:22.022 19:04:39 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:22.022 ************************************ 00:14:22.022 START TEST raid5f_state_function_test 00:14:22.022 ************************************ 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 4 false 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=90854 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:22.022 Process raid pid: 90854 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 90854' 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 90854 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # '[' -z 90854 ']' 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:22.022 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:22.022 19:04:39 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:22.022 [2024-12-05 19:04:39.455104] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:14:22.022 [2024-12-05 19:04:39.455242] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:22.283 [2024-12-05 19:04:39.611764] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:22.283 [2024-12-05 19:04:39.636791] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:22.283 [2024-12-05 19:04:39.680091] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:22.283 [2024-12-05 19:04:39.680129] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@868 -- # return 0 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:22.853 [2024-12-05 19:04:40.275368] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:22.853 [2024-12-05 19:04:40.275429] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:22.853 [2024-12-05 19:04:40.275446] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:22.853 [2024-12-05 19:04:40.275456] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:22.853 [2024-12-05 19:04:40.275462] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:22.853 [2024-12-05 19:04:40.275474] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:22.853 [2024-12-05 19:04:40.275480] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:22.853 [2024-12-05 19:04:40.275488] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:22.853 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:22.853 "name": "Existed_Raid", 00:14:22.853 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:22.853 "strip_size_kb": 64, 00:14:22.853 "state": "configuring", 00:14:22.853 "raid_level": "raid5f", 00:14:22.853 "superblock": false, 00:14:22.853 "num_base_bdevs": 4, 00:14:22.853 "num_base_bdevs_discovered": 0, 00:14:22.854 "num_base_bdevs_operational": 4, 00:14:22.854 "base_bdevs_list": [ 00:14:22.854 { 00:14:22.854 "name": "BaseBdev1", 00:14:22.854 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:22.854 "is_configured": false, 00:14:22.854 "data_offset": 0, 00:14:22.854 "data_size": 0 00:14:22.854 }, 00:14:22.854 { 00:14:22.854 "name": "BaseBdev2", 00:14:22.854 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:22.854 "is_configured": false, 00:14:22.854 "data_offset": 0, 00:14:22.854 "data_size": 0 00:14:22.854 }, 00:14:22.854 { 00:14:22.854 "name": "BaseBdev3", 00:14:22.854 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:22.854 "is_configured": false, 00:14:22.854 "data_offset": 0, 00:14:22.854 "data_size": 0 00:14:22.854 }, 00:14:22.854 { 00:14:22.854 "name": "BaseBdev4", 00:14:22.854 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:22.854 "is_configured": false, 00:14:22.854 "data_offset": 0, 00:14:22.854 "data_size": 0 00:14:22.854 } 00:14:22.854 ] 00:14:22.854 }' 00:14:22.854 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:22.854 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.425 [2024-12-05 19:04:40.678619] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:23.425 [2024-12-05 19:04:40.678659] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.425 [2024-12-05 19:04:40.690620] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:23.425 [2024-12-05 19:04:40.690660] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:23.425 [2024-12-05 19:04:40.690678] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:23.425 [2024-12-05 19:04:40.690686] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:23.425 [2024-12-05 19:04:40.690692] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:23.425 [2024-12-05 19:04:40.690700] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:23.425 [2024-12-05 19:04:40.690705] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:23.425 [2024-12-05 19:04:40.690713] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.425 [2024-12-05 19:04:40.711432] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:23.425 BaseBdev1 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:23.425 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.426 [ 00:14:23.426 { 00:14:23.426 "name": "BaseBdev1", 00:14:23.426 "aliases": [ 00:14:23.426 "14e7f018-e6f7-4f65-b280-a778d94a6d47" 00:14:23.426 ], 00:14:23.426 "product_name": "Malloc disk", 00:14:23.426 "block_size": 512, 00:14:23.426 "num_blocks": 65536, 00:14:23.426 "uuid": "14e7f018-e6f7-4f65-b280-a778d94a6d47", 00:14:23.426 "assigned_rate_limits": { 00:14:23.426 "rw_ios_per_sec": 0, 00:14:23.426 "rw_mbytes_per_sec": 0, 00:14:23.426 "r_mbytes_per_sec": 0, 00:14:23.426 "w_mbytes_per_sec": 0 00:14:23.426 }, 00:14:23.426 "claimed": true, 00:14:23.426 "claim_type": "exclusive_write", 00:14:23.426 "zoned": false, 00:14:23.426 "supported_io_types": { 00:14:23.426 "read": true, 00:14:23.426 "write": true, 00:14:23.426 "unmap": true, 00:14:23.426 "flush": true, 00:14:23.426 "reset": true, 00:14:23.426 "nvme_admin": false, 00:14:23.426 "nvme_io": false, 00:14:23.426 "nvme_io_md": false, 00:14:23.426 "write_zeroes": true, 00:14:23.426 "zcopy": true, 00:14:23.426 "get_zone_info": false, 00:14:23.426 "zone_management": false, 00:14:23.426 "zone_append": false, 00:14:23.426 "compare": false, 00:14:23.426 "compare_and_write": false, 00:14:23.426 "abort": true, 00:14:23.426 "seek_hole": false, 00:14:23.426 "seek_data": false, 00:14:23.426 "copy": true, 00:14:23.426 "nvme_iov_md": false 00:14:23.426 }, 00:14:23.426 "memory_domains": [ 00:14:23.426 { 00:14:23.426 "dma_device_id": "system", 00:14:23.426 "dma_device_type": 1 00:14:23.426 }, 00:14:23.426 { 00:14:23.426 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:23.426 "dma_device_type": 2 00:14:23.426 } 00:14:23.426 ], 00:14:23.426 "driver_specific": {} 00:14:23.426 } 00:14:23.426 ] 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:23.426 "name": "Existed_Raid", 00:14:23.426 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:23.426 "strip_size_kb": 64, 00:14:23.426 "state": "configuring", 00:14:23.426 "raid_level": "raid5f", 00:14:23.426 "superblock": false, 00:14:23.426 "num_base_bdevs": 4, 00:14:23.426 "num_base_bdevs_discovered": 1, 00:14:23.426 "num_base_bdevs_operational": 4, 00:14:23.426 "base_bdevs_list": [ 00:14:23.426 { 00:14:23.426 "name": "BaseBdev1", 00:14:23.426 "uuid": "14e7f018-e6f7-4f65-b280-a778d94a6d47", 00:14:23.426 "is_configured": true, 00:14:23.426 "data_offset": 0, 00:14:23.426 "data_size": 65536 00:14:23.426 }, 00:14:23.426 { 00:14:23.426 "name": "BaseBdev2", 00:14:23.426 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:23.426 "is_configured": false, 00:14:23.426 "data_offset": 0, 00:14:23.426 "data_size": 0 00:14:23.426 }, 00:14:23.426 { 00:14:23.426 "name": "BaseBdev3", 00:14:23.426 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:23.426 "is_configured": false, 00:14:23.426 "data_offset": 0, 00:14:23.426 "data_size": 0 00:14:23.426 }, 00:14:23.426 { 00:14:23.426 "name": "BaseBdev4", 00:14:23.426 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:23.426 "is_configured": false, 00:14:23.426 "data_offset": 0, 00:14:23.426 "data_size": 0 00:14:23.426 } 00:14:23.426 ] 00:14:23.426 }' 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:23.426 19:04:40 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.714 [2024-12-05 19:04:41.206603] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:23.714 [2024-12-05 19:04:41.206645] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.714 [2024-12-05 19:04:41.218614] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:23.714 [2024-12-05 19:04:41.220419] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:23.714 [2024-12-05 19:04:41.220461] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:23.714 [2024-12-05 19:04:41.220470] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:23.714 [2024-12-05 19:04:41.220478] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:23.714 [2024-12-05 19:04:41.220483] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:23.714 [2024-12-05 19:04:41.220491] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.714 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.004 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:24.004 "name": "Existed_Raid", 00:14:24.004 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:24.004 "strip_size_kb": 64, 00:14:24.004 "state": "configuring", 00:14:24.004 "raid_level": "raid5f", 00:14:24.004 "superblock": false, 00:14:24.004 "num_base_bdevs": 4, 00:14:24.004 "num_base_bdevs_discovered": 1, 00:14:24.004 "num_base_bdevs_operational": 4, 00:14:24.004 "base_bdevs_list": [ 00:14:24.004 { 00:14:24.004 "name": "BaseBdev1", 00:14:24.004 "uuid": "14e7f018-e6f7-4f65-b280-a778d94a6d47", 00:14:24.004 "is_configured": true, 00:14:24.004 "data_offset": 0, 00:14:24.004 "data_size": 65536 00:14:24.004 }, 00:14:24.004 { 00:14:24.004 "name": "BaseBdev2", 00:14:24.004 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:24.004 "is_configured": false, 00:14:24.004 "data_offset": 0, 00:14:24.004 "data_size": 0 00:14:24.004 }, 00:14:24.004 { 00:14:24.004 "name": "BaseBdev3", 00:14:24.004 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:24.004 "is_configured": false, 00:14:24.004 "data_offset": 0, 00:14:24.004 "data_size": 0 00:14:24.004 }, 00:14:24.004 { 00:14:24.004 "name": "BaseBdev4", 00:14:24.004 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:24.004 "is_configured": false, 00:14:24.004 "data_offset": 0, 00:14:24.004 "data_size": 0 00:14:24.004 } 00:14:24.004 ] 00:14:24.004 }' 00:14:24.004 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:24.004 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.281 [2024-12-05 19:04:41.728644] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:24.281 BaseBdev2 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.281 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.281 [ 00:14:24.281 { 00:14:24.281 "name": "BaseBdev2", 00:14:24.281 "aliases": [ 00:14:24.281 "86f1d197-bd68-474f-8fe5-781fcf8b9cc5" 00:14:24.281 ], 00:14:24.281 "product_name": "Malloc disk", 00:14:24.281 "block_size": 512, 00:14:24.281 "num_blocks": 65536, 00:14:24.281 "uuid": "86f1d197-bd68-474f-8fe5-781fcf8b9cc5", 00:14:24.281 "assigned_rate_limits": { 00:14:24.281 "rw_ios_per_sec": 0, 00:14:24.281 "rw_mbytes_per_sec": 0, 00:14:24.281 "r_mbytes_per_sec": 0, 00:14:24.281 "w_mbytes_per_sec": 0 00:14:24.281 }, 00:14:24.281 "claimed": true, 00:14:24.281 "claim_type": "exclusive_write", 00:14:24.281 "zoned": false, 00:14:24.281 "supported_io_types": { 00:14:24.281 "read": true, 00:14:24.281 "write": true, 00:14:24.281 "unmap": true, 00:14:24.281 "flush": true, 00:14:24.281 "reset": true, 00:14:24.281 "nvme_admin": false, 00:14:24.281 "nvme_io": false, 00:14:24.281 "nvme_io_md": false, 00:14:24.281 "write_zeroes": true, 00:14:24.281 "zcopy": true, 00:14:24.281 "get_zone_info": false, 00:14:24.281 "zone_management": false, 00:14:24.281 "zone_append": false, 00:14:24.281 "compare": false, 00:14:24.281 "compare_and_write": false, 00:14:24.281 "abort": true, 00:14:24.281 "seek_hole": false, 00:14:24.281 "seek_data": false, 00:14:24.281 "copy": true, 00:14:24.281 "nvme_iov_md": false 00:14:24.281 }, 00:14:24.281 "memory_domains": [ 00:14:24.281 { 00:14:24.282 "dma_device_id": "system", 00:14:24.282 "dma_device_type": 1 00:14:24.282 }, 00:14:24.282 { 00:14:24.282 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:24.282 "dma_device_type": 2 00:14:24.282 } 00:14:24.282 ], 00:14:24.282 "driver_specific": {} 00:14:24.282 } 00:14:24.282 ] 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:24.282 "name": "Existed_Raid", 00:14:24.282 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:24.282 "strip_size_kb": 64, 00:14:24.282 "state": "configuring", 00:14:24.282 "raid_level": "raid5f", 00:14:24.282 "superblock": false, 00:14:24.282 "num_base_bdevs": 4, 00:14:24.282 "num_base_bdevs_discovered": 2, 00:14:24.282 "num_base_bdevs_operational": 4, 00:14:24.282 "base_bdevs_list": [ 00:14:24.282 { 00:14:24.282 "name": "BaseBdev1", 00:14:24.282 "uuid": "14e7f018-e6f7-4f65-b280-a778d94a6d47", 00:14:24.282 "is_configured": true, 00:14:24.282 "data_offset": 0, 00:14:24.282 "data_size": 65536 00:14:24.282 }, 00:14:24.282 { 00:14:24.282 "name": "BaseBdev2", 00:14:24.282 "uuid": "86f1d197-bd68-474f-8fe5-781fcf8b9cc5", 00:14:24.282 "is_configured": true, 00:14:24.282 "data_offset": 0, 00:14:24.282 "data_size": 65536 00:14:24.282 }, 00:14:24.282 { 00:14:24.282 "name": "BaseBdev3", 00:14:24.282 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:24.282 "is_configured": false, 00:14:24.282 "data_offset": 0, 00:14:24.282 "data_size": 0 00:14:24.282 }, 00:14:24.282 { 00:14:24.282 "name": "BaseBdev4", 00:14:24.282 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:24.282 "is_configured": false, 00:14:24.282 "data_offset": 0, 00:14:24.282 "data_size": 0 00:14:24.282 } 00:14:24.282 ] 00:14:24.282 }' 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:24.282 19:04:41 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.852 [2024-12-05 19:04:42.228299] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:24.852 BaseBdev3 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.852 [ 00:14:24.852 { 00:14:24.852 "name": "BaseBdev3", 00:14:24.852 "aliases": [ 00:14:24.852 "029308cd-3c60-4e48-ba8c-174f673aa621" 00:14:24.852 ], 00:14:24.852 "product_name": "Malloc disk", 00:14:24.852 "block_size": 512, 00:14:24.852 "num_blocks": 65536, 00:14:24.852 "uuid": "029308cd-3c60-4e48-ba8c-174f673aa621", 00:14:24.852 "assigned_rate_limits": { 00:14:24.852 "rw_ios_per_sec": 0, 00:14:24.852 "rw_mbytes_per_sec": 0, 00:14:24.852 "r_mbytes_per_sec": 0, 00:14:24.852 "w_mbytes_per_sec": 0 00:14:24.852 }, 00:14:24.852 "claimed": true, 00:14:24.852 "claim_type": "exclusive_write", 00:14:24.852 "zoned": false, 00:14:24.852 "supported_io_types": { 00:14:24.852 "read": true, 00:14:24.852 "write": true, 00:14:24.852 "unmap": true, 00:14:24.852 "flush": true, 00:14:24.852 "reset": true, 00:14:24.852 "nvme_admin": false, 00:14:24.852 "nvme_io": false, 00:14:24.852 "nvme_io_md": false, 00:14:24.852 "write_zeroes": true, 00:14:24.852 "zcopy": true, 00:14:24.852 "get_zone_info": false, 00:14:24.852 "zone_management": false, 00:14:24.852 "zone_append": false, 00:14:24.852 "compare": false, 00:14:24.852 "compare_and_write": false, 00:14:24.852 "abort": true, 00:14:24.852 "seek_hole": false, 00:14:24.852 "seek_data": false, 00:14:24.852 "copy": true, 00:14:24.852 "nvme_iov_md": false 00:14:24.852 }, 00:14:24.852 "memory_domains": [ 00:14:24.852 { 00:14:24.852 "dma_device_id": "system", 00:14:24.852 "dma_device_type": 1 00:14:24.852 }, 00:14:24.852 { 00:14:24.852 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:24.852 "dma_device_type": 2 00:14:24.852 } 00:14:24.852 ], 00:14:24.852 "driver_specific": {} 00:14:24.852 } 00:14:24.852 ] 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:24.852 "name": "Existed_Raid", 00:14:24.852 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:24.852 "strip_size_kb": 64, 00:14:24.852 "state": "configuring", 00:14:24.852 "raid_level": "raid5f", 00:14:24.852 "superblock": false, 00:14:24.852 "num_base_bdevs": 4, 00:14:24.852 "num_base_bdevs_discovered": 3, 00:14:24.852 "num_base_bdevs_operational": 4, 00:14:24.852 "base_bdevs_list": [ 00:14:24.852 { 00:14:24.852 "name": "BaseBdev1", 00:14:24.852 "uuid": "14e7f018-e6f7-4f65-b280-a778d94a6d47", 00:14:24.852 "is_configured": true, 00:14:24.852 "data_offset": 0, 00:14:24.852 "data_size": 65536 00:14:24.852 }, 00:14:24.852 { 00:14:24.852 "name": "BaseBdev2", 00:14:24.852 "uuid": "86f1d197-bd68-474f-8fe5-781fcf8b9cc5", 00:14:24.852 "is_configured": true, 00:14:24.852 "data_offset": 0, 00:14:24.852 "data_size": 65536 00:14:24.852 }, 00:14:24.852 { 00:14:24.852 "name": "BaseBdev3", 00:14:24.852 "uuid": "029308cd-3c60-4e48-ba8c-174f673aa621", 00:14:24.852 "is_configured": true, 00:14:24.852 "data_offset": 0, 00:14:24.852 "data_size": 65536 00:14:24.852 }, 00:14:24.852 { 00:14:24.852 "name": "BaseBdev4", 00:14:24.852 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:24.852 "is_configured": false, 00:14:24.852 "data_offset": 0, 00:14:24.852 "data_size": 0 00:14:24.852 } 00:14:24.852 ] 00:14:24.852 }' 00:14:24.852 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:24.853 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:25.423 [2024-12-05 19:04:42.722532] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:25.423 [2024-12-05 19:04:42.722585] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:25.423 [2024-12-05 19:04:42.722594] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:25.423 [2024-12-05 19:04:42.722885] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:25.423 BaseBdev4 00:14:25.423 [2024-12-05 19:04:42.723347] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:25.423 [2024-12-05 19:04:42.723378] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:14:25.423 [2024-12-05 19:04:42.723558] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:25.423 [ 00:14:25.423 { 00:14:25.423 "name": "BaseBdev4", 00:14:25.423 "aliases": [ 00:14:25.423 "8b2ccbcb-daed-4910-88ea-370f2c92a161" 00:14:25.423 ], 00:14:25.423 "product_name": "Malloc disk", 00:14:25.423 "block_size": 512, 00:14:25.423 "num_blocks": 65536, 00:14:25.423 "uuid": "8b2ccbcb-daed-4910-88ea-370f2c92a161", 00:14:25.423 "assigned_rate_limits": { 00:14:25.423 "rw_ios_per_sec": 0, 00:14:25.423 "rw_mbytes_per_sec": 0, 00:14:25.423 "r_mbytes_per_sec": 0, 00:14:25.423 "w_mbytes_per_sec": 0 00:14:25.423 }, 00:14:25.423 "claimed": true, 00:14:25.423 "claim_type": "exclusive_write", 00:14:25.423 "zoned": false, 00:14:25.423 "supported_io_types": { 00:14:25.423 "read": true, 00:14:25.423 "write": true, 00:14:25.423 "unmap": true, 00:14:25.423 "flush": true, 00:14:25.423 "reset": true, 00:14:25.423 "nvme_admin": false, 00:14:25.423 "nvme_io": false, 00:14:25.423 "nvme_io_md": false, 00:14:25.423 "write_zeroes": true, 00:14:25.423 "zcopy": true, 00:14:25.423 "get_zone_info": false, 00:14:25.423 "zone_management": false, 00:14:25.423 "zone_append": false, 00:14:25.423 "compare": false, 00:14:25.423 "compare_and_write": false, 00:14:25.423 "abort": true, 00:14:25.423 "seek_hole": false, 00:14:25.423 "seek_data": false, 00:14:25.423 "copy": true, 00:14:25.423 "nvme_iov_md": false 00:14:25.423 }, 00:14:25.423 "memory_domains": [ 00:14:25.423 { 00:14:25.423 "dma_device_id": "system", 00:14:25.423 "dma_device_type": 1 00:14:25.423 }, 00:14:25.423 { 00:14:25.423 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:25.423 "dma_device_type": 2 00:14:25.423 } 00:14:25.423 ], 00:14:25.423 "driver_specific": {} 00:14:25.423 } 00:14:25.423 ] 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:25.423 "name": "Existed_Raid", 00:14:25.423 "uuid": "be3712b0-1a83-4f35-b93e-09cbc26b9d1b", 00:14:25.423 "strip_size_kb": 64, 00:14:25.423 "state": "online", 00:14:25.423 "raid_level": "raid5f", 00:14:25.423 "superblock": false, 00:14:25.423 "num_base_bdevs": 4, 00:14:25.423 "num_base_bdevs_discovered": 4, 00:14:25.423 "num_base_bdevs_operational": 4, 00:14:25.423 "base_bdevs_list": [ 00:14:25.423 { 00:14:25.423 "name": "BaseBdev1", 00:14:25.423 "uuid": "14e7f018-e6f7-4f65-b280-a778d94a6d47", 00:14:25.423 "is_configured": true, 00:14:25.423 "data_offset": 0, 00:14:25.423 "data_size": 65536 00:14:25.423 }, 00:14:25.423 { 00:14:25.423 "name": "BaseBdev2", 00:14:25.423 "uuid": "86f1d197-bd68-474f-8fe5-781fcf8b9cc5", 00:14:25.423 "is_configured": true, 00:14:25.423 "data_offset": 0, 00:14:25.423 "data_size": 65536 00:14:25.423 }, 00:14:25.423 { 00:14:25.423 "name": "BaseBdev3", 00:14:25.423 "uuid": "029308cd-3c60-4e48-ba8c-174f673aa621", 00:14:25.423 "is_configured": true, 00:14:25.423 "data_offset": 0, 00:14:25.423 "data_size": 65536 00:14:25.423 }, 00:14:25.423 { 00:14:25.423 "name": "BaseBdev4", 00:14:25.423 "uuid": "8b2ccbcb-daed-4910-88ea-370f2c92a161", 00:14:25.423 "is_configured": true, 00:14:25.423 "data_offset": 0, 00:14:25.423 "data_size": 65536 00:14:25.423 } 00:14:25.423 ] 00:14:25.423 }' 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:25.423 19:04:42 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:25.683 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:25.684 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:25.684 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:25.684 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:25.684 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:25.684 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:25.684 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:25.684 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:25.684 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.684 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:25.684 [2024-12-05 19:04:43.201915] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:25.684 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.684 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:25.684 "name": "Existed_Raid", 00:14:25.684 "aliases": [ 00:14:25.684 "be3712b0-1a83-4f35-b93e-09cbc26b9d1b" 00:14:25.684 ], 00:14:25.684 "product_name": "Raid Volume", 00:14:25.684 "block_size": 512, 00:14:25.684 "num_blocks": 196608, 00:14:25.684 "uuid": "be3712b0-1a83-4f35-b93e-09cbc26b9d1b", 00:14:25.684 "assigned_rate_limits": { 00:14:25.684 "rw_ios_per_sec": 0, 00:14:25.684 "rw_mbytes_per_sec": 0, 00:14:25.684 "r_mbytes_per_sec": 0, 00:14:25.684 "w_mbytes_per_sec": 0 00:14:25.684 }, 00:14:25.684 "claimed": false, 00:14:25.684 "zoned": false, 00:14:25.684 "supported_io_types": { 00:14:25.684 "read": true, 00:14:25.684 "write": true, 00:14:25.684 "unmap": false, 00:14:25.684 "flush": false, 00:14:25.684 "reset": true, 00:14:25.684 "nvme_admin": false, 00:14:25.684 "nvme_io": false, 00:14:25.684 "nvme_io_md": false, 00:14:25.684 "write_zeroes": true, 00:14:25.684 "zcopy": false, 00:14:25.684 "get_zone_info": false, 00:14:25.684 "zone_management": false, 00:14:25.684 "zone_append": false, 00:14:25.684 "compare": false, 00:14:25.684 "compare_and_write": false, 00:14:25.684 "abort": false, 00:14:25.684 "seek_hole": false, 00:14:25.684 "seek_data": false, 00:14:25.684 "copy": false, 00:14:25.684 "nvme_iov_md": false 00:14:25.684 }, 00:14:25.684 "driver_specific": { 00:14:25.684 "raid": { 00:14:25.684 "uuid": "be3712b0-1a83-4f35-b93e-09cbc26b9d1b", 00:14:25.684 "strip_size_kb": 64, 00:14:25.684 "state": "online", 00:14:25.684 "raid_level": "raid5f", 00:14:25.684 "superblock": false, 00:14:25.684 "num_base_bdevs": 4, 00:14:25.684 "num_base_bdevs_discovered": 4, 00:14:25.684 "num_base_bdevs_operational": 4, 00:14:25.684 "base_bdevs_list": [ 00:14:25.684 { 00:14:25.684 "name": "BaseBdev1", 00:14:25.684 "uuid": "14e7f018-e6f7-4f65-b280-a778d94a6d47", 00:14:25.684 "is_configured": true, 00:14:25.684 "data_offset": 0, 00:14:25.684 "data_size": 65536 00:14:25.684 }, 00:14:25.684 { 00:14:25.684 "name": "BaseBdev2", 00:14:25.684 "uuid": "86f1d197-bd68-474f-8fe5-781fcf8b9cc5", 00:14:25.684 "is_configured": true, 00:14:25.684 "data_offset": 0, 00:14:25.684 "data_size": 65536 00:14:25.684 }, 00:14:25.684 { 00:14:25.684 "name": "BaseBdev3", 00:14:25.684 "uuid": "029308cd-3c60-4e48-ba8c-174f673aa621", 00:14:25.684 "is_configured": true, 00:14:25.684 "data_offset": 0, 00:14:25.684 "data_size": 65536 00:14:25.684 }, 00:14:25.684 { 00:14:25.684 "name": "BaseBdev4", 00:14:25.684 "uuid": "8b2ccbcb-daed-4910-88ea-370f2c92a161", 00:14:25.684 "is_configured": true, 00:14:25.684 "data_offset": 0, 00:14:25.684 "data_size": 65536 00:14:25.684 } 00:14:25.684 ] 00:14:25.684 } 00:14:25.684 } 00:14:25.684 }' 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:25.944 BaseBdev2 00:14:25.944 BaseBdev3 00:14:25.944 BaseBdev4' 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:25.944 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.204 [2024-12-05 19:04:43.521223] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:26.204 "name": "Existed_Raid", 00:14:26.204 "uuid": "be3712b0-1a83-4f35-b93e-09cbc26b9d1b", 00:14:26.204 "strip_size_kb": 64, 00:14:26.204 "state": "online", 00:14:26.204 "raid_level": "raid5f", 00:14:26.204 "superblock": false, 00:14:26.204 "num_base_bdevs": 4, 00:14:26.204 "num_base_bdevs_discovered": 3, 00:14:26.204 "num_base_bdevs_operational": 3, 00:14:26.204 "base_bdevs_list": [ 00:14:26.204 { 00:14:26.204 "name": null, 00:14:26.204 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:26.204 "is_configured": false, 00:14:26.204 "data_offset": 0, 00:14:26.204 "data_size": 65536 00:14:26.204 }, 00:14:26.204 { 00:14:26.204 "name": "BaseBdev2", 00:14:26.204 "uuid": "86f1d197-bd68-474f-8fe5-781fcf8b9cc5", 00:14:26.204 "is_configured": true, 00:14:26.204 "data_offset": 0, 00:14:26.204 "data_size": 65536 00:14:26.204 }, 00:14:26.204 { 00:14:26.204 "name": "BaseBdev3", 00:14:26.204 "uuid": "029308cd-3c60-4e48-ba8c-174f673aa621", 00:14:26.204 "is_configured": true, 00:14:26.204 "data_offset": 0, 00:14:26.204 "data_size": 65536 00:14:26.204 }, 00:14:26.204 { 00:14:26.204 "name": "BaseBdev4", 00:14:26.204 "uuid": "8b2ccbcb-daed-4910-88ea-370f2c92a161", 00:14:26.204 "is_configured": true, 00:14:26.204 "data_offset": 0, 00:14:26.204 "data_size": 65536 00:14:26.204 } 00:14:26.204 ] 00:14:26.204 }' 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:26.204 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.464 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:26.464 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:26.464 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.464 19:04:43 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:26.464 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.464 19:04:43 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.464 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.723 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.724 [2024-12-05 19:04:44.039802] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:26.724 [2024-12-05 19:04:44.039978] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:26.724 [2024-12-05 19:04:44.050922] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.724 [2024-12-05 19:04:44.110848] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.724 [2024-12-05 19:04:44.181536] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:14:26.724 [2024-12-05 19:04:44.181625] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.724 BaseBdev2 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.724 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.985 [ 00:14:26.985 { 00:14:26.985 "name": "BaseBdev2", 00:14:26.985 "aliases": [ 00:14:26.985 "d24b1f6a-b5d2-4c63-b4b6-dadecdb25dad" 00:14:26.985 ], 00:14:26.985 "product_name": "Malloc disk", 00:14:26.985 "block_size": 512, 00:14:26.985 "num_blocks": 65536, 00:14:26.985 "uuid": "d24b1f6a-b5d2-4c63-b4b6-dadecdb25dad", 00:14:26.985 "assigned_rate_limits": { 00:14:26.985 "rw_ios_per_sec": 0, 00:14:26.985 "rw_mbytes_per_sec": 0, 00:14:26.985 "r_mbytes_per_sec": 0, 00:14:26.985 "w_mbytes_per_sec": 0 00:14:26.985 }, 00:14:26.985 "claimed": false, 00:14:26.985 "zoned": false, 00:14:26.985 "supported_io_types": { 00:14:26.985 "read": true, 00:14:26.985 "write": true, 00:14:26.985 "unmap": true, 00:14:26.985 "flush": true, 00:14:26.985 "reset": true, 00:14:26.985 "nvme_admin": false, 00:14:26.985 "nvme_io": false, 00:14:26.985 "nvme_io_md": false, 00:14:26.985 "write_zeroes": true, 00:14:26.985 "zcopy": true, 00:14:26.985 "get_zone_info": false, 00:14:26.985 "zone_management": false, 00:14:26.985 "zone_append": false, 00:14:26.985 "compare": false, 00:14:26.985 "compare_and_write": false, 00:14:26.985 "abort": true, 00:14:26.985 "seek_hole": false, 00:14:26.985 "seek_data": false, 00:14:26.985 "copy": true, 00:14:26.985 "nvme_iov_md": false 00:14:26.985 }, 00:14:26.985 "memory_domains": [ 00:14:26.985 { 00:14:26.985 "dma_device_id": "system", 00:14:26.985 "dma_device_type": 1 00:14:26.985 }, 00:14:26.985 { 00:14:26.985 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:26.985 "dma_device_type": 2 00:14:26.985 } 00:14:26.985 ], 00:14:26.985 "driver_specific": {} 00:14:26.985 } 00:14:26.985 ] 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.985 BaseBdev3 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.985 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.985 [ 00:14:26.985 { 00:14:26.985 "name": "BaseBdev3", 00:14:26.985 "aliases": [ 00:14:26.985 "830e2b63-0dd9-4b6a-9eb0-1cbc415d0f1f" 00:14:26.985 ], 00:14:26.985 "product_name": "Malloc disk", 00:14:26.985 "block_size": 512, 00:14:26.985 "num_blocks": 65536, 00:14:26.985 "uuid": "830e2b63-0dd9-4b6a-9eb0-1cbc415d0f1f", 00:14:26.985 "assigned_rate_limits": { 00:14:26.985 "rw_ios_per_sec": 0, 00:14:26.985 "rw_mbytes_per_sec": 0, 00:14:26.985 "r_mbytes_per_sec": 0, 00:14:26.985 "w_mbytes_per_sec": 0 00:14:26.985 }, 00:14:26.985 "claimed": false, 00:14:26.985 "zoned": false, 00:14:26.985 "supported_io_types": { 00:14:26.985 "read": true, 00:14:26.985 "write": true, 00:14:26.985 "unmap": true, 00:14:26.985 "flush": true, 00:14:26.985 "reset": true, 00:14:26.985 "nvme_admin": false, 00:14:26.985 "nvme_io": false, 00:14:26.985 "nvme_io_md": false, 00:14:26.985 "write_zeroes": true, 00:14:26.985 "zcopy": true, 00:14:26.985 "get_zone_info": false, 00:14:26.985 "zone_management": false, 00:14:26.985 "zone_append": false, 00:14:26.985 "compare": false, 00:14:26.985 "compare_and_write": false, 00:14:26.985 "abort": true, 00:14:26.985 "seek_hole": false, 00:14:26.985 "seek_data": false, 00:14:26.985 "copy": true, 00:14:26.985 "nvme_iov_md": false 00:14:26.985 }, 00:14:26.985 "memory_domains": [ 00:14:26.985 { 00:14:26.985 "dma_device_id": "system", 00:14:26.985 "dma_device_type": 1 00:14:26.986 }, 00:14:26.986 { 00:14:26.986 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:26.986 "dma_device_type": 2 00:14:26.986 } 00:14:26.986 ], 00:14:26.986 "driver_specific": {} 00:14:26.986 } 00:14:26.986 ] 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.986 BaseBdev4 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.986 [ 00:14:26.986 { 00:14:26.986 "name": "BaseBdev4", 00:14:26.986 "aliases": [ 00:14:26.986 "7717cf5d-a92d-4c40-8b24-8fedf074a314" 00:14:26.986 ], 00:14:26.986 "product_name": "Malloc disk", 00:14:26.986 "block_size": 512, 00:14:26.986 "num_blocks": 65536, 00:14:26.986 "uuid": "7717cf5d-a92d-4c40-8b24-8fedf074a314", 00:14:26.986 "assigned_rate_limits": { 00:14:26.986 "rw_ios_per_sec": 0, 00:14:26.986 "rw_mbytes_per_sec": 0, 00:14:26.986 "r_mbytes_per_sec": 0, 00:14:26.986 "w_mbytes_per_sec": 0 00:14:26.986 }, 00:14:26.986 "claimed": false, 00:14:26.986 "zoned": false, 00:14:26.986 "supported_io_types": { 00:14:26.986 "read": true, 00:14:26.986 "write": true, 00:14:26.986 "unmap": true, 00:14:26.986 "flush": true, 00:14:26.986 "reset": true, 00:14:26.986 "nvme_admin": false, 00:14:26.986 "nvme_io": false, 00:14:26.986 "nvme_io_md": false, 00:14:26.986 "write_zeroes": true, 00:14:26.986 "zcopy": true, 00:14:26.986 "get_zone_info": false, 00:14:26.986 "zone_management": false, 00:14:26.986 "zone_append": false, 00:14:26.986 "compare": false, 00:14:26.986 "compare_and_write": false, 00:14:26.986 "abort": true, 00:14:26.986 "seek_hole": false, 00:14:26.986 "seek_data": false, 00:14:26.986 "copy": true, 00:14:26.986 "nvme_iov_md": false 00:14:26.986 }, 00:14:26.986 "memory_domains": [ 00:14:26.986 { 00:14:26.986 "dma_device_id": "system", 00:14:26.986 "dma_device_type": 1 00:14:26.986 }, 00:14:26.986 { 00:14:26.986 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:26.986 "dma_device_type": 2 00:14:26.986 } 00:14:26.986 ], 00:14:26.986 "driver_specific": {} 00:14:26.986 } 00:14:26.986 ] 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.986 [2024-12-05 19:04:44.404416] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:26.986 [2024-12-05 19:04:44.404535] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:26.986 [2024-12-05 19:04:44.404582] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:26.986 [2024-12-05 19:04:44.406420] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:26.986 [2024-12-05 19:04:44.406521] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:26.986 "name": "Existed_Raid", 00:14:26.986 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:26.986 "strip_size_kb": 64, 00:14:26.986 "state": "configuring", 00:14:26.986 "raid_level": "raid5f", 00:14:26.986 "superblock": false, 00:14:26.986 "num_base_bdevs": 4, 00:14:26.986 "num_base_bdevs_discovered": 3, 00:14:26.986 "num_base_bdevs_operational": 4, 00:14:26.986 "base_bdevs_list": [ 00:14:26.986 { 00:14:26.986 "name": "BaseBdev1", 00:14:26.986 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:26.986 "is_configured": false, 00:14:26.986 "data_offset": 0, 00:14:26.986 "data_size": 0 00:14:26.986 }, 00:14:26.986 { 00:14:26.986 "name": "BaseBdev2", 00:14:26.986 "uuid": "d24b1f6a-b5d2-4c63-b4b6-dadecdb25dad", 00:14:26.986 "is_configured": true, 00:14:26.986 "data_offset": 0, 00:14:26.986 "data_size": 65536 00:14:26.986 }, 00:14:26.986 { 00:14:26.986 "name": "BaseBdev3", 00:14:26.986 "uuid": "830e2b63-0dd9-4b6a-9eb0-1cbc415d0f1f", 00:14:26.986 "is_configured": true, 00:14:26.986 "data_offset": 0, 00:14:26.986 "data_size": 65536 00:14:26.986 }, 00:14:26.986 { 00:14:26.986 "name": "BaseBdev4", 00:14:26.986 "uuid": "7717cf5d-a92d-4c40-8b24-8fedf074a314", 00:14:26.986 "is_configured": true, 00:14:26.986 "data_offset": 0, 00:14:26.986 "data_size": 65536 00:14:26.986 } 00:14:26.986 ] 00:14:26.986 }' 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:26.986 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:27.557 [2024-12-05 19:04:44.819710] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:27.557 "name": "Existed_Raid", 00:14:27.557 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:27.557 "strip_size_kb": 64, 00:14:27.557 "state": "configuring", 00:14:27.557 "raid_level": "raid5f", 00:14:27.557 "superblock": false, 00:14:27.557 "num_base_bdevs": 4, 00:14:27.557 "num_base_bdevs_discovered": 2, 00:14:27.557 "num_base_bdevs_operational": 4, 00:14:27.557 "base_bdevs_list": [ 00:14:27.557 { 00:14:27.557 "name": "BaseBdev1", 00:14:27.557 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:27.557 "is_configured": false, 00:14:27.557 "data_offset": 0, 00:14:27.557 "data_size": 0 00:14:27.557 }, 00:14:27.557 { 00:14:27.557 "name": null, 00:14:27.557 "uuid": "d24b1f6a-b5d2-4c63-b4b6-dadecdb25dad", 00:14:27.557 "is_configured": false, 00:14:27.557 "data_offset": 0, 00:14:27.557 "data_size": 65536 00:14:27.557 }, 00:14:27.557 { 00:14:27.557 "name": "BaseBdev3", 00:14:27.557 "uuid": "830e2b63-0dd9-4b6a-9eb0-1cbc415d0f1f", 00:14:27.557 "is_configured": true, 00:14:27.557 "data_offset": 0, 00:14:27.557 "data_size": 65536 00:14:27.557 }, 00:14:27.557 { 00:14:27.557 "name": "BaseBdev4", 00:14:27.557 "uuid": "7717cf5d-a92d-4c40-8b24-8fedf074a314", 00:14:27.557 "is_configured": true, 00:14:27.557 "data_offset": 0, 00:14:27.557 "data_size": 65536 00:14:27.557 } 00:14:27.557 ] 00:14:27.557 }' 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:27.557 19:04:44 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:27.817 [2024-12-05 19:04:45.309782] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:27.817 BaseBdev1 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:27.817 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:27.818 [ 00:14:27.818 { 00:14:27.818 "name": "BaseBdev1", 00:14:27.818 "aliases": [ 00:14:27.818 "7ddaac15-69d4-4ef3-afde-f552ba0001bd" 00:14:27.818 ], 00:14:27.818 "product_name": "Malloc disk", 00:14:27.818 "block_size": 512, 00:14:27.818 "num_blocks": 65536, 00:14:27.818 "uuid": "7ddaac15-69d4-4ef3-afde-f552ba0001bd", 00:14:27.818 "assigned_rate_limits": { 00:14:27.818 "rw_ios_per_sec": 0, 00:14:27.818 "rw_mbytes_per_sec": 0, 00:14:27.818 "r_mbytes_per_sec": 0, 00:14:27.818 "w_mbytes_per_sec": 0 00:14:27.818 }, 00:14:27.818 "claimed": true, 00:14:27.818 "claim_type": "exclusive_write", 00:14:27.818 "zoned": false, 00:14:27.818 "supported_io_types": { 00:14:27.818 "read": true, 00:14:27.818 "write": true, 00:14:27.818 "unmap": true, 00:14:27.818 "flush": true, 00:14:27.818 "reset": true, 00:14:27.818 "nvme_admin": false, 00:14:27.818 "nvme_io": false, 00:14:27.818 "nvme_io_md": false, 00:14:27.818 "write_zeroes": true, 00:14:27.818 "zcopy": true, 00:14:27.818 "get_zone_info": false, 00:14:27.818 "zone_management": false, 00:14:27.818 "zone_append": false, 00:14:27.818 "compare": false, 00:14:27.818 "compare_and_write": false, 00:14:27.818 "abort": true, 00:14:27.818 "seek_hole": false, 00:14:27.818 "seek_data": false, 00:14:27.818 "copy": true, 00:14:27.818 "nvme_iov_md": false 00:14:27.818 }, 00:14:27.818 "memory_domains": [ 00:14:27.818 { 00:14:27.818 "dma_device_id": "system", 00:14:27.818 "dma_device_type": 1 00:14:27.818 }, 00:14:27.818 { 00:14:27.818 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:27.818 "dma_device_type": 2 00:14:27.818 } 00:14:27.818 ], 00:14:27.818 "driver_specific": {} 00:14:27.818 } 00:14:27.818 ] 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:27.818 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.077 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:28.077 "name": "Existed_Raid", 00:14:28.077 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:28.077 "strip_size_kb": 64, 00:14:28.077 "state": "configuring", 00:14:28.077 "raid_level": "raid5f", 00:14:28.077 "superblock": false, 00:14:28.077 "num_base_bdevs": 4, 00:14:28.077 "num_base_bdevs_discovered": 3, 00:14:28.077 "num_base_bdevs_operational": 4, 00:14:28.077 "base_bdevs_list": [ 00:14:28.077 { 00:14:28.077 "name": "BaseBdev1", 00:14:28.077 "uuid": "7ddaac15-69d4-4ef3-afde-f552ba0001bd", 00:14:28.077 "is_configured": true, 00:14:28.077 "data_offset": 0, 00:14:28.077 "data_size": 65536 00:14:28.077 }, 00:14:28.077 { 00:14:28.077 "name": null, 00:14:28.077 "uuid": "d24b1f6a-b5d2-4c63-b4b6-dadecdb25dad", 00:14:28.077 "is_configured": false, 00:14:28.077 "data_offset": 0, 00:14:28.077 "data_size": 65536 00:14:28.077 }, 00:14:28.077 { 00:14:28.077 "name": "BaseBdev3", 00:14:28.077 "uuid": "830e2b63-0dd9-4b6a-9eb0-1cbc415d0f1f", 00:14:28.077 "is_configured": true, 00:14:28.077 "data_offset": 0, 00:14:28.077 "data_size": 65536 00:14:28.077 }, 00:14:28.077 { 00:14:28.077 "name": "BaseBdev4", 00:14:28.077 "uuid": "7717cf5d-a92d-4c40-8b24-8fedf074a314", 00:14:28.077 "is_configured": true, 00:14:28.077 "data_offset": 0, 00:14:28.077 "data_size": 65536 00:14:28.077 } 00:14:28.077 ] 00:14:28.077 }' 00:14:28.077 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:28.077 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.337 [2024-12-05 19:04:45.840921] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.337 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.598 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:28.598 "name": "Existed_Raid", 00:14:28.598 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:28.598 "strip_size_kb": 64, 00:14:28.598 "state": "configuring", 00:14:28.598 "raid_level": "raid5f", 00:14:28.598 "superblock": false, 00:14:28.598 "num_base_bdevs": 4, 00:14:28.598 "num_base_bdevs_discovered": 2, 00:14:28.598 "num_base_bdevs_operational": 4, 00:14:28.598 "base_bdevs_list": [ 00:14:28.598 { 00:14:28.598 "name": "BaseBdev1", 00:14:28.598 "uuid": "7ddaac15-69d4-4ef3-afde-f552ba0001bd", 00:14:28.598 "is_configured": true, 00:14:28.598 "data_offset": 0, 00:14:28.598 "data_size": 65536 00:14:28.598 }, 00:14:28.598 { 00:14:28.598 "name": null, 00:14:28.598 "uuid": "d24b1f6a-b5d2-4c63-b4b6-dadecdb25dad", 00:14:28.598 "is_configured": false, 00:14:28.598 "data_offset": 0, 00:14:28.598 "data_size": 65536 00:14:28.598 }, 00:14:28.598 { 00:14:28.598 "name": null, 00:14:28.598 "uuid": "830e2b63-0dd9-4b6a-9eb0-1cbc415d0f1f", 00:14:28.598 "is_configured": false, 00:14:28.598 "data_offset": 0, 00:14:28.598 "data_size": 65536 00:14:28.598 }, 00:14:28.598 { 00:14:28.598 "name": "BaseBdev4", 00:14:28.598 "uuid": "7717cf5d-a92d-4c40-8b24-8fedf074a314", 00:14:28.598 "is_configured": true, 00:14:28.598 "data_offset": 0, 00:14:28.598 "data_size": 65536 00:14:28.598 } 00:14:28.598 ] 00:14:28.598 }' 00:14:28.598 19:04:45 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:28.598 19:04:45 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.857 [2024-12-05 19:04:46.372060] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.857 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.117 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:29.117 "name": "Existed_Raid", 00:14:29.117 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:29.117 "strip_size_kb": 64, 00:14:29.117 "state": "configuring", 00:14:29.117 "raid_level": "raid5f", 00:14:29.117 "superblock": false, 00:14:29.117 "num_base_bdevs": 4, 00:14:29.117 "num_base_bdevs_discovered": 3, 00:14:29.117 "num_base_bdevs_operational": 4, 00:14:29.117 "base_bdevs_list": [ 00:14:29.117 { 00:14:29.117 "name": "BaseBdev1", 00:14:29.117 "uuid": "7ddaac15-69d4-4ef3-afde-f552ba0001bd", 00:14:29.117 "is_configured": true, 00:14:29.117 "data_offset": 0, 00:14:29.117 "data_size": 65536 00:14:29.117 }, 00:14:29.117 { 00:14:29.117 "name": null, 00:14:29.117 "uuid": "d24b1f6a-b5d2-4c63-b4b6-dadecdb25dad", 00:14:29.117 "is_configured": false, 00:14:29.117 "data_offset": 0, 00:14:29.117 "data_size": 65536 00:14:29.117 }, 00:14:29.117 { 00:14:29.117 "name": "BaseBdev3", 00:14:29.117 "uuid": "830e2b63-0dd9-4b6a-9eb0-1cbc415d0f1f", 00:14:29.117 "is_configured": true, 00:14:29.117 "data_offset": 0, 00:14:29.117 "data_size": 65536 00:14:29.117 }, 00:14:29.117 { 00:14:29.117 "name": "BaseBdev4", 00:14:29.117 "uuid": "7717cf5d-a92d-4c40-8b24-8fedf074a314", 00:14:29.117 "is_configured": true, 00:14:29.117 "data_offset": 0, 00:14:29.117 "data_size": 65536 00:14:29.117 } 00:14:29.117 ] 00:14:29.117 }' 00:14:29.117 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:29.117 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.377 [2024-12-05 19:04:46.847366] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.377 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:29.377 "name": "Existed_Raid", 00:14:29.377 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:29.377 "strip_size_kb": 64, 00:14:29.377 "state": "configuring", 00:14:29.377 "raid_level": "raid5f", 00:14:29.377 "superblock": false, 00:14:29.377 "num_base_bdevs": 4, 00:14:29.377 "num_base_bdevs_discovered": 2, 00:14:29.377 "num_base_bdevs_operational": 4, 00:14:29.377 "base_bdevs_list": [ 00:14:29.377 { 00:14:29.377 "name": null, 00:14:29.378 "uuid": "7ddaac15-69d4-4ef3-afde-f552ba0001bd", 00:14:29.378 "is_configured": false, 00:14:29.378 "data_offset": 0, 00:14:29.378 "data_size": 65536 00:14:29.378 }, 00:14:29.378 { 00:14:29.378 "name": null, 00:14:29.378 "uuid": "d24b1f6a-b5d2-4c63-b4b6-dadecdb25dad", 00:14:29.378 "is_configured": false, 00:14:29.378 "data_offset": 0, 00:14:29.378 "data_size": 65536 00:14:29.378 }, 00:14:29.378 { 00:14:29.378 "name": "BaseBdev3", 00:14:29.378 "uuid": "830e2b63-0dd9-4b6a-9eb0-1cbc415d0f1f", 00:14:29.378 "is_configured": true, 00:14:29.378 "data_offset": 0, 00:14:29.378 "data_size": 65536 00:14:29.378 }, 00:14:29.378 { 00:14:29.378 "name": "BaseBdev4", 00:14:29.378 "uuid": "7717cf5d-a92d-4c40-8b24-8fedf074a314", 00:14:29.378 "is_configured": true, 00:14:29.378 "data_offset": 0, 00:14:29.378 "data_size": 65536 00:14:29.378 } 00:14:29.378 ] 00:14:29.378 }' 00:14:29.378 19:04:46 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:29.378 19:04:46 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.948 [2024-12-05 19:04:47.393028] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:29.948 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:29.948 "name": "Existed_Raid", 00:14:29.948 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:29.948 "strip_size_kb": 64, 00:14:29.948 "state": "configuring", 00:14:29.948 "raid_level": "raid5f", 00:14:29.948 "superblock": false, 00:14:29.948 "num_base_bdevs": 4, 00:14:29.948 "num_base_bdevs_discovered": 3, 00:14:29.948 "num_base_bdevs_operational": 4, 00:14:29.948 "base_bdevs_list": [ 00:14:29.948 { 00:14:29.948 "name": null, 00:14:29.948 "uuid": "7ddaac15-69d4-4ef3-afde-f552ba0001bd", 00:14:29.948 "is_configured": false, 00:14:29.948 "data_offset": 0, 00:14:29.948 "data_size": 65536 00:14:29.948 }, 00:14:29.948 { 00:14:29.948 "name": "BaseBdev2", 00:14:29.948 "uuid": "d24b1f6a-b5d2-4c63-b4b6-dadecdb25dad", 00:14:29.948 "is_configured": true, 00:14:29.949 "data_offset": 0, 00:14:29.949 "data_size": 65536 00:14:29.949 }, 00:14:29.949 { 00:14:29.949 "name": "BaseBdev3", 00:14:29.949 "uuid": "830e2b63-0dd9-4b6a-9eb0-1cbc415d0f1f", 00:14:29.949 "is_configured": true, 00:14:29.949 "data_offset": 0, 00:14:29.949 "data_size": 65536 00:14:29.949 }, 00:14:29.949 { 00:14:29.949 "name": "BaseBdev4", 00:14:29.949 "uuid": "7717cf5d-a92d-4c40-8b24-8fedf074a314", 00:14:29.949 "is_configured": true, 00:14:29.949 "data_offset": 0, 00:14:29.949 "data_size": 65536 00:14:29.949 } 00:14:29.949 ] 00:14:29.949 }' 00:14:29.949 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:29.949 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.519 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:30.519 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:30.519 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:30.519 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.519 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:30.519 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:30.519 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:30.519 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:30.519 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:30.519 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.519 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:30.519 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 7ddaac15-69d4-4ef3-afde-f552ba0001bd 00:14:30.519 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:30.519 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.520 [2024-12-05 19:04:47.910685] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:30.520 [2024-12-05 19:04:47.910789] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:30.520 [2024-12-05 19:04:47.910812] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:30.520 [2024-12-05 19:04:47.911095] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:14:30.520 [2024-12-05 19:04:47.911564] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:30.520 [2024-12-05 19:04:47.911615] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:14:30.520 [2024-12-05 19:04:47.911825] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:30.520 NewBaseBdev 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@905 -- # local i 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.520 [ 00:14:30.520 { 00:14:30.520 "name": "NewBaseBdev", 00:14:30.520 "aliases": [ 00:14:30.520 "7ddaac15-69d4-4ef3-afde-f552ba0001bd" 00:14:30.520 ], 00:14:30.520 "product_name": "Malloc disk", 00:14:30.520 "block_size": 512, 00:14:30.520 "num_blocks": 65536, 00:14:30.520 "uuid": "7ddaac15-69d4-4ef3-afde-f552ba0001bd", 00:14:30.520 "assigned_rate_limits": { 00:14:30.520 "rw_ios_per_sec": 0, 00:14:30.520 "rw_mbytes_per_sec": 0, 00:14:30.520 "r_mbytes_per_sec": 0, 00:14:30.520 "w_mbytes_per_sec": 0 00:14:30.520 }, 00:14:30.520 "claimed": true, 00:14:30.520 "claim_type": "exclusive_write", 00:14:30.520 "zoned": false, 00:14:30.520 "supported_io_types": { 00:14:30.520 "read": true, 00:14:30.520 "write": true, 00:14:30.520 "unmap": true, 00:14:30.520 "flush": true, 00:14:30.520 "reset": true, 00:14:30.520 "nvme_admin": false, 00:14:30.520 "nvme_io": false, 00:14:30.520 "nvme_io_md": false, 00:14:30.520 "write_zeroes": true, 00:14:30.520 "zcopy": true, 00:14:30.520 "get_zone_info": false, 00:14:30.520 "zone_management": false, 00:14:30.520 "zone_append": false, 00:14:30.520 "compare": false, 00:14:30.520 "compare_and_write": false, 00:14:30.520 "abort": true, 00:14:30.520 "seek_hole": false, 00:14:30.520 "seek_data": false, 00:14:30.520 "copy": true, 00:14:30.520 "nvme_iov_md": false 00:14:30.520 }, 00:14:30.520 "memory_domains": [ 00:14:30.520 { 00:14:30.520 "dma_device_id": "system", 00:14:30.520 "dma_device_type": 1 00:14:30.520 }, 00:14:30.520 { 00:14:30.520 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:30.520 "dma_device_type": 2 00:14:30.520 } 00:14:30.520 ], 00:14:30.520 "driver_specific": {} 00:14:30.520 } 00:14:30.520 ] 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@911 -- # return 0 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:30.520 "name": "Existed_Raid", 00:14:30.520 "uuid": "17df839f-9e60-41d3-98cf-42355599f7b9", 00:14:30.520 "strip_size_kb": 64, 00:14:30.520 "state": "online", 00:14:30.520 "raid_level": "raid5f", 00:14:30.520 "superblock": false, 00:14:30.520 "num_base_bdevs": 4, 00:14:30.520 "num_base_bdevs_discovered": 4, 00:14:30.520 "num_base_bdevs_operational": 4, 00:14:30.520 "base_bdevs_list": [ 00:14:30.520 { 00:14:30.520 "name": "NewBaseBdev", 00:14:30.520 "uuid": "7ddaac15-69d4-4ef3-afde-f552ba0001bd", 00:14:30.520 "is_configured": true, 00:14:30.520 "data_offset": 0, 00:14:30.520 "data_size": 65536 00:14:30.520 }, 00:14:30.520 { 00:14:30.520 "name": "BaseBdev2", 00:14:30.520 "uuid": "d24b1f6a-b5d2-4c63-b4b6-dadecdb25dad", 00:14:30.520 "is_configured": true, 00:14:30.520 "data_offset": 0, 00:14:30.520 "data_size": 65536 00:14:30.520 }, 00:14:30.520 { 00:14:30.520 "name": "BaseBdev3", 00:14:30.520 "uuid": "830e2b63-0dd9-4b6a-9eb0-1cbc415d0f1f", 00:14:30.520 "is_configured": true, 00:14:30.520 "data_offset": 0, 00:14:30.520 "data_size": 65536 00:14:30.520 }, 00:14:30.520 { 00:14:30.520 "name": "BaseBdev4", 00:14:30.520 "uuid": "7717cf5d-a92d-4c40-8b24-8fedf074a314", 00:14:30.520 "is_configured": true, 00:14:30.520 "data_offset": 0, 00:14:30.520 "data_size": 65536 00:14:30.520 } 00:14:30.520 ] 00:14:30.520 }' 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:30.520 19:04:47 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.090 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:31.090 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:31.090 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:31.090 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:31.090 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.091 [2024-12-05 19:04:48.421972] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:31.091 "name": "Existed_Raid", 00:14:31.091 "aliases": [ 00:14:31.091 "17df839f-9e60-41d3-98cf-42355599f7b9" 00:14:31.091 ], 00:14:31.091 "product_name": "Raid Volume", 00:14:31.091 "block_size": 512, 00:14:31.091 "num_blocks": 196608, 00:14:31.091 "uuid": "17df839f-9e60-41d3-98cf-42355599f7b9", 00:14:31.091 "assigned_rate_limits": { 00:14:31.091 "rw_ios_per_sec": 0, 00:14:31.091 "rw_mbytes_per_sec": 0, 00:14:31.091 "r_mbytes_per_sec": 0, 00:14:31.091 "w_mbytes_per_sec": 0 00:14:31.091 }, 00:14:31.091 "claimed": false, 00:14:31.091 "zoned": false, 00:14:31.091 "supported_io_types": { 00:14:31.091 "read": true, 00:14:31.091 "write": true, 00:14:31.091 "unmap": false, 00:14:31.091 "flush": false, 00:14:31.091 "reset": true, 00:14:31.091 "nvme_admin": false, 00:14:31.091 "nvme_io": false, 00:14:31.091 "nvme_io_md": false, 00:14:31.091 "write_zeroes": true, 00:14:31.091 "zcopy": false, 00:14:31.091 "get_zone_info": false, 00:14:31.091 "zone_management": false, 00:14:31.091 "zone_append": false, 00:14:31.091 "compare": false, 00:14:31.091 "compare_and_write": false, 00:14:31.091 "abort": false, 00:14:31.091 "seek_hole": false, 00:14:31.091 "seek_data": false, 00:14:31.091 "copy": false, 00:14:31.091 "nvme_iov_md": false 00:14:31.091 }, 00:14:31.091 "driver_specific": { 00:14:31.091 "raid": { 00:14:31.091 "uuid": "17df839f-9e60-41d3-98cf-42355599f7b9", 00:14:31.091 "strip_size_kb": 64, 00:14:31.091 "state": "online", 00:14:31.091 "raid_level": "raid5f", 00:14:31.091 "superblock": false, 00:14:31.091 "num_base_bdevs": 4, 00:14:31.091 "num_base_bdevs_discovered": 4, 00:14:31.091 "num_base_bdevs_operational": 4, 00:14:31.091 "base_bdevs_list": [ 00:14:31.091 { 00:14:31.091 "name": "NewBaseBdev", 00:14:31.091 "uuid": "7ddaac15-69d4-4ef3-afde-f552ba0001bd", 00:14:31.091 "is_configured": true, 00:14:31.091 "data_offset": 0, 00:14:31.091 "data_size": 65536 00:14:31.091 }, 00:14:31.091 { 00:14:31.091 "name": "BaseBdev2", 00:14:31.091 "uuid": "d24b1f6a-b5d2-4c63-b4b6-dadecdb25dad", 00:14:31.091 "is_configured": true, 00:14:31.091 "data_offset": 0, 00:14:31.091 "data_size": 65536 00:14:31.091 }, 00:14:31.091 { 00:14:31.091 "name": "BaseBdev3", 00:14:31.091 "uuid": "830e2b63-0dd9-4b6a-9eb0-1cbc415d0f1f", 00:14:31.091 "is_configured": true, 00:14:31.091 "data_offset": 0, 00:14:31.091 "data_size": 65536 00:14:31.091 }, 00:14:31.091 { 00:14:31.091 "name": "BaseBdev4", 00:14:31.091 "uuid": "7717cf5d-a92d-4c40-8b24-8fedf074a314", 00:14:31.091 "is_configured": true, 00:14:31.091 "data_offset": 0, 00:14:31.091 "data_size": 65536 00:14:31.091 } 00:14:31.091 ] 00:14:31.091 } 00:14:31.091 } 00:14:31.091 }' 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:31.091 BaseBdev2 00:14:31.091 BaseBdev3 00:14:31.091 BaseBdev4' 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:31.091 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.352 [2024-12-05 19:04:48.737305] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:31.352 [2024-12-05 19:04:48.737379] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:31.352 [2024-12-05 19:04:48.737482] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:31.352 [2024-12-05 19:04:48.737766] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:31.352 [2024-12-05 19:04:48.737820] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 90854 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # '[' -z 90854 ']' 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@958 -- # kill -0 90854 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # uname 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 90854 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:31.352 killing process with pid 90854 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 90854' 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@973 -- # kill 90854 00:14:31.352 [2024-12-05 19:04:48.786543] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:31.352 19:04:48 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@978 -- # wait 90854 00:14:31.352 [2024-12-05 19:04:48.826778] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:31.613 19:04:49 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:14:31.613 00:14:31.613 real 0m9.687s 00:14:31.613 user 0m16.584s 00:14:31.613 sys 0m2.083s 00:14:31.613 ************************************ 00:14:31.613 END TEST raid5f_state_function_test 00:14:31.613 ************************************ 00:14:31.613 19:04:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:31.613 19:04:49 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.613 19:04:49 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 4 true 00:14:31.613 19:04:49 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:14:31.613 19:04:49 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:31.613 19:04:49 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:31.613 ************************************ 00:14:31.613 START TEST raid5f_state_function_test_sb 00:14:31.614 ************************************ 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1129 -- # raid_state_function_test raid5f 4 true 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=91449 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 91449' 00:14:31.614 Process raid pid: 91449 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 91449 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # '[' -z 91449 ']' 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:31.614 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:31.614 19:04:49 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:31.874 [2024-12-05 19:04:49.238529] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:14:31.874 [2024-12-05 19:04:49.238720] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:31.874 [2024-12-05 19:04:49.397140] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:31.874 [2024-12-05 19:04:49.421833] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:32.134 [2024-12-05 19:04:49.465551] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:32.134 [2024-12-05 19:04:49.465586] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:32.704 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:32.704 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@868 -- # return 0 00:14:32.704 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:32.704 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:32.704 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:32.704 [2024-12-05 19:04:50.072881] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:32.704 [2024-12-05 19:04:50.073002] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:32.704 [2024-12-05 19:04:50.073060] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:32.704 [2024-12-05 19:04:50.073084] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:32.704 [2024-12-05 19:04:50.073101] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:32.704 [2024-12-05 19:04:50.073122] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:32.704 [2024-12-05 19:04:50.073139] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:32.704 [2024-12-05 19:04:50.073158] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:32.704 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:32.704 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:32.704 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:32.704 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:32.704 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:32.704 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:32.704 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:32.704 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:32.704 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:32.704 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:32.705 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:32.705 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:32.705 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:32.705 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:32.705 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:32.705 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:32.705 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:32.705 "name": "Existed_Raid", 00:14:32.705 "uuid": "37222218-640b-4faa-ab3b-a7be6206d046", 00:14:32.705 "strip_size_kb": 64, 00:14:32.705 "state": "configuring", 00:14:32.705 "raid_level": "raid5f", 00:14:32.705 "superblock": true, 00:14:32.705 "num_base_bdevs": 4, 00:14:32.705 "num_base_bdevs_discovered": 0, 00:14:32.705 "num_base_bdevs_operational": 4, 00:14:32.705 "base_bdevs_list": [ 00:14:32.705 { 00:14:32.705 "name": "BaseBdev1", 00:14:32.705 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:32.705 "is_configured": false, 00:14:32.705 "data_offset": 0, 00:14:32.705 "data_size": 0 00:14:32.705 }, 00:14:32.705 { 00:14:32.705 "name": "BaseBdev2", 00:14:32.705 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:32.705 "is_configured": false, 00:14:32.705 "data_offset": 0, 00:14:32.705 "data_size": 0 00:14:32.705 }, 00:14:32.705 { 00:14:32.705 "name": "BaseBdev3", 00:14:32.705 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:32.705 "is_configured": false, 00:14:32.705 "data_offset": 0, 00:14:32.705 "data_size": 0 00:14:32.705 }, 00:14:32.705 { 00:14:32.705 "name": "BaseBdev4", 00:14:32.705 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:32.705 "is_configured": false, 00:14:32.705 "data_offset": 0, 00:14:32.705 "data_size": 0 00:14:32.705 } 00:14:32.705 ] 00:14:32.705 }' 00:14:32.705 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:32.705 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.275 [2024-12-05 19:04:50.536027] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:33.275 [2024-12-05 19:04:50.536106] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.275 [2024-12-05 19:04:50.548032] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:33.275 [2024-12-05 19:04:50.548111] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:33.275 [2024-12-05 19:04:50.548151] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:33.275 [2024-12-05 19:04:50.548172] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:33.275 [2024-12-05 19:04:50.548189] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:33.275 [2024-12-05 19:04:50.548208] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:33.275 [2024-12-05 19:04:50.548224] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:33.275 [2024-12-05 19:04:50.548243] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.275 [2024-12-05 19:04:50.568898] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:33.275 BaseBdev1 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.275 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.275 [ 00:14:33.275 { 00:14:33.275 "name": "BaseBdev1", 00:14:33.275 "aliases": [ 00:14:33.275 "663442b2-0227-4550-9aeb-431bea4b2e8c" 00:14:33.275 ], 00:14:33.275 "product_name": "Malloc disk", 00:14:33.275 "block_size": 512, 00:14:33.275 "num_blocks": 65536, 00:14:33.275 "uuid": "663442b2-0227-4550-9aeb-431bea4b2e8c", 00:14:33.275 "assigned_rate_limits": { 00:14:33.275 "rw_ios_per_sec": 0, 00:14:33.275 "rw_mbytes_per_sec": 0, 00:14:33.275 "r_mbytes_per_sec": 0, 00:14:33.276 "w_mbytes_per_sec": 0 00:14:33.276 }, 00:14:33.276 "claimed": true, 00:14:33.276 "claim_type": "exclusive_write", 00:14:33.276 "zoned": false, 00:14:33.276 "supported_io_types": { 00:14:33.276 "read": true, 00:14:33.276 "write": true, 00:14:33.276 "unmap": true, 00:14:33.276 "flush": true, 00:14:33.276 "reset": true, 00:14:33.276 "nvme_admin": false, 00:14:33.276 "nvme_io": false, 00:14:33.276 "nvme_io_md": false, 00:14:33.276 "write_zeroes": true, 00:14:33.276 "zcopy": true, 00:14:33.276 "get_zone_info": false, 00:14:33.276 "zone_management": false, 00:14:33.276 "zone_append": false, 00:14:33.276 "compare": false, 00:14:33.276 "compare_and_write": false, 00:14:33.276 "abort": true, 00:14:33.276 "seek_hole": false, 00:14:33.276 "seek_data": false, 00:14:33.276 "copy": true, 00:14:33.276 "nvme_iov_md": false 00:14:33.276 }, 00:14:33.276 "memory_domains": [ 00:14:33.276 { 00:14:33.276 "dma_device_id": "system", 00:14:33.276 "dma_device_type": 1 00:14:33.276 }, 00:14:33.276 { 00:14:33.276 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:33.276 "dma_device_type": 2 00:14:33.276 } 00:14:33.276 ], 00:14:33.276 "driver_specific": {} 00:14:33.276 } 00:14:33.276 ] 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:33.276 "name": "Existed_Raid", 00:14:33.276 "uuid": "2abf6dca-2e73-4674-808d-34efe3478286", 00:14:33.276 "strip_size_kb": 64, 00:14:33.276 "state": "configuring", 00:14:33.276 "raid_level": "raid5f", 00:14:33.276 "superblock": true, 00:14:33.276 "num_base_bdevs": 4, 00:14:33.276 "num_base_bdevs_discovered": 1, 00:14:33.276 "num_base_bdevs_operational": 4, 00:14:33.276 "base_bdevs_list": [ 00:14:33.276 { 00:14:33.276 "name": "BaseBdev1", 00:14:33.276 "uuid": "663442b2-0227-4550-9aeb-431bea4b2e8c", 00:14:33.276 "is_configured": true, 00:14:33.276 "data_offset": 2048, 00:14:33.276 "data_size": 63488 00:14:33.276 }, 00:14:33.276 { 00:14:33.276 "name": "BaseBdev2", 00:14:33.276 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.276 "is_configured": false, 00:14:33.276 "data_offset": 0, 00:14:33.276 "data_size": 0 00:14:33.276 }, 00:14:33.276 { 00:14:33.276 "name": "BaseBdev3", 00:14:33.276 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.276 "is_configured": false, 00:14:33.276 "data_offset": 0, 00:14:33.276 "data_size": 0 00:14:33.276 }, 00:14:33.276 { 00:14:33.276 "name": "BaseBdev4", 00:14:33.276 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.276 "is_configured": false, 00:14:33.276 "data_offset": 0, 00:14:33.276 "data_size": 0 00:14:33.276 } 00:14:33.276 ] 00:14:33.276 }' 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:33.276 19:04:50 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.535 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:33.535 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.535 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.535 [2024-12-05 19:04:51.036123] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:33.535 [2024-12-05 19:04:51.036212] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:14:33.535 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.535 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:33.535 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.535 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.535 [2024-12-05 19:04:51.048138] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:33.535 [2024-12-05 19:04:51.050063] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:33.535 [2024-12-05 19:04:51.050134] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:33.535 [2024-12-05 19:04:51.050187] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:33.535 [2024-12-05 19:04:51.050208] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:33.535 [2024-12-05 19:04:51.050225] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:33.535 [2024-12-05 19:04:51.050244] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:33.535 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.535 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:33.535 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:33.535 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:33.536 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:33.536 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:33.536 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:33.536 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:33.536 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:33.536 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:33.536 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:33.536 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:33.536 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:33.536 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.536 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:33.536 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:33.536 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.536 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:33.796 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:33.796 "name": "Existed_Raid", 00:14:33.796 "uuid": "3bb29c02-0c83-4dd4-80f1-fe8cc0c49222", 00:14:33.796 "strip_size_kb": 64, 00:14:33.796 "state": "configuring", 00:14:33.796 "raid_level": "raid5f", 00:14:33.796 "superblock": true, 00:14:33.796 "num_base_bdevs": 4, 00:14:33.796 "num_base_bdevs_discovered": 1, 00:14:33.796 "num_base_bdevs_operational": 4, 00:14:33.796 "base_bdevs_list": [ 00:14:33.796 { 00:14:33.796 "name": "BaseBdev1", 00:14:33.796 "uuid": "663442b2-0227-4550-9aeb-431bea4b2e8c", 00:14:33.796 "is_configured": true, 00:14:33.796 "data_offset": 2048, 00:14:33.796 "data_size": 63488 00:14:33.796 }, 00:14:33.796 { 00:14:33.796 "name": "BaseBdev2", 00:14:33.796 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.796 "is_configured": false, 00:14:33.796 "data_offset": 0, 00:14:33.796 "data_size": 0 00:14:33.796 }, 00:14:33.796 { 00:14:33.796 "name": "BaseBdev3", 00:14:33.796 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.796 "is_configured": false, 00:14:33.796 "data_offset": 0, 00:14:33.796 "data_size": 0 00:14:33.796 }, 00:14:33.796 { 00:14:33.796 "name": "BaseBdev4", 00:14:33.796 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.796 "is_configured": false, 00:14:33.796 "data_offset": 0, 00:14:33.796 "data_size": 0 00:14:33.796 } 00:14:33.796 ] 00:14:33.796 }' 00:14:33.796 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:33.796 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:34.056 [2024-12-05 19:04:51.482638] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:34.056 BaseBdev2 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:34.056 [ 00:14:34.056 { 00:14:34.056 "name": "BaseBdev2", 00:14:34.056 "aliases": [ 00:14:34.056 "122672ca-f7a4-4366-80ea-c7da78ee7ae5" 00:14:34.056 ], 00:14:34.056 "product_name": "Malloc disk", 00:14:34.056 "block_size": 512, 00:14:34.056 "num_blocks": 65536, 00:14:34.056 "uuid": "122672ca-f7a4-4366-80ea-c7da78ee7ae5", 00:14:34.056 "assigned_rate_limits": { 00:14:34.056 "rw_ios_per_sec": 0, 00:14:34.056 "rw_mbytes_per_sec": 0, 00:14:34.056 "r_mbytes_per_sec": 0, 00:14:34.056 "w_mbytes_per_sec": 0 00:14:34.056 }, 00:14:34.056 "claimed": true, 00:14:34.056 "claim_type": "exclusive_write", 00:14:34.056 "zoned": false, 00:14:34.056 "supported_io_types": { 00:14:34.056 "read": true, 00:14:34.056 "write": true, 00:14:34.056 "unmap": true, 00:14:34.056 "flush": true, 00:14:34.056 "reset": true, 00:14:34.056 "nvme_admin": false, 00:14:34.056 "nvme_io": false, 00:14:34.056 "nvme_io_md": false, 00:14:34.056 "write_zeroes": true, 00:14:34.056 "zcopy": true, 00:14:34.056 "get_zone_info": false, 00:14:34.056 "zone_management": false, 00:14:34.056 "zone_append": false, 00:14:34.056 "compare": false, 00:14:34.056 "compare_and_write": false, 00:14:34.056 "abort": true, 00:14:34.056 "seek_hole": false, 00:14:34.056 "seek_data": false, 00:14:34.056 "copy": true, 00:14:34.056 "nvme_iov_md": false 00:14:34.056 }, 00:14:34.056 "memory_domains": [ 00:14:34.056 { 00:14:34.056 "dma_device_id": "system", 00:14:34.056 "dma_device_type": 1 00:14:34.056 }, 00:14:34.056 { 00:14:34.056 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:34.056 "dma_device_type": 2 00:14:34.056 } 00:14:34.056 ], 00:14:34.056 "driver_specific": {} 00:14:34.056 } 00:14:34.056 ] 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:34.056 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.057 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:34.057 "name": "Existed_Raid", 00:14:34.057 "uuid": "3bb29c02-0c83-4dd4-80f1-fe8cc0c49222", 00:14:34.057 "strip_size_kb": 64, 00:14:34.057 "state": "configuring", 00:14:34.057 "raid_level": "raid5f", 00:14:34.057 "superblock": true, 00:14:34.057 "num_base_bdevs": 4, 00:14:34.057 "num_base_bdevs_discovered": 2, 00:14:34.057 "num_base_bdevs_operational": 4, 00:14:34.057 "base_bdevs_list": [ 00:14:34.057 { 00:14:34.057 "name": "BaseBdev1", 00:14:34.057 "uuid": "663442b2-0227-4550-9aeb-431bea4b2e8c", 00:14:34.057 "is_configured": true, 00:14:34.057 "data_offset": 2048, 00:14:34.057 "data_size": 63488 00:14:34.057 }, 00:14:34.057 { 00:14:34.057 "name": "BaseBdev2", 00:14:34.057 "uuid": "122672ca-f7a4-4366-80ea-c7da78ee7ae5", 00:14:34.057 "is_configured": true, 00:14:34.057 "data_offset": 2048, 00:14:34.057 "data_size": 63488 00:14:34.057 }, 00:14:34.057 { 00:14:34.057 "name": "BaseBdev3", 00:14:34.057 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:34.057 "is_configured": false, 00:14:34.057 "data_offset": 0, 00:14:34.057 "data_size": 0 00:14:34.057 }, 00:14:34.057 { 00:14:34.057 "name": "BaseBdev4", 00:14:34.057 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:34.057 "is_configured": false, 00:14:34.057 "data_offset": 0, 00:14:34.057 "data_size": 0 00:14:34.057 } 00:14:34.057 ] 00:14:34.057 }' 00:14:34.057 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:34.057 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:34.627 [2024-12-05 19:04:51.949218] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:34.627 BaseBdev3 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:34.627 [ 00:14:34.627 { 00:14:34.627 "name": "BaseBdev3", 00:14:34.627 "aliases": [ 00:14:34.627 "060b1088-d8db-4c57-8fd5-5e0fc40f0ccd" 00:14:34.627 ], 00:14:34.627 "product_name": "Malloc disk", 00:14:34.627 "block_size": 512, 00:14:34.627 "num_blocks": 65536, 00:14:34.627 "uuid": "060b1088-d8db-4c57-8fd5-5e0fc40f0ccd", 00:14:34.627 "assigned_rate_limits": { 00:14:34.627 "rw_ios_per_sec": 0, 00:14:34.627 "rw_mbytes_per_sec": 0, 00:14:34.627 "r_mbytes_per_sec": 0, 00:14:34.627 "w_mbytes_per_sec": 0 00:14:34.627 }, 00:14:34.627 "claimed": true, 00:14:34.627 "claim_type": "exclusive_write", 00:14:34.627 "zoned": false, 00:14:34.627 "supported_io_types": { 00:14:34.627 "read": true, 00:14:34.627 "write": true, 00:14:34.627 "unmap": true, 00:14:34.627 "flush": true, 00:14:34.627 "reset": true, 00:14:34.627 "nvme_admin": false, 00:14:34.627 "nvme_io": false, 00:14:34.627 "nvme_io_md": false, 00:14:34.627 "write_zeroes": true, 00:14:34.627 "zcopy": true, 00:14:34.627 "get_zone_info": false, 00:14:34.627 "zone_management": false, 00:14:34.627 "zone_append": false, 00:14:34.627 "compare": false, 00:14:34.627 "compare_and_write": false, 00:14:34.627 "abort": true, 00:14:34.627 "seek_hole": false, 00:14:34.627 "seek_data": false, 00:14:34.627 "copy": true, 00:14:34.627 "nvme_iov_md": false 00:14:34.627 }, 00:14:34.627 "memory_domains": [ 00:14:34.627 { 00:14:34.627 "dma_device_id": "system", 00:14:34.627 "dma_device_type": 1 00:14:34.627 }, 00:14:34.627 { 00:14:34.627 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:34.627 "dma_device_type": 2 00:14:34.627 } 00:14:34.627 ], 00:14:34.627 "driver_specific": {} 00:14:34.627 } 00:14:34.627 ] 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:34.627 19:04:51 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:34.627 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:34.627 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:34.627 "name": "Existed_Raid", 00:14:34.627 "uuid": "3bb29c02-0c83-4dd4-80f1-fe8cc0c49222", 00:14:34.627 "strip_size_kb": 64, 00:14:34.627 "state": "configuring", 00:14:34.628 "raid_level": "raid5f", 00:14:34.628 "superblock": true, 00:14:34.628 "num_base_bdevs": 4, 00:14:34.628 "num_base_bdevs_discovered": 3, 00:14:34.628 "num_base_bdevs_operational": 4, 00:14:34.628 "base_bdevs_list": [ 00:14:34.628 { 00:14:34.628 "name": "BaseBdev1", 00:14:34.628 "uuid": "663442b2-0227-4550-9aeb-431bea4b2e8c", 00:14:34.628 "is_configured": true, 00:14:34.628 "data_offset": 2048, 00:14:34.628 "data_size": 63488 00:14:34.628 }, 00:14:34.628 { 00:14:34.628 "name": "BaseBdev2", 00:14:34.628 "uuid": "122672ca-f7a4-4366-80ea-c7da78ee7ae5", 00:14:34.628 "is_configured": true, 00:14:34.628 "data_offset": 2048, 00:14:34.628 "data_size": 63488 00:14:34.628 }, 00:14:34.628 { 00:14:34.628 "name": "BaseBdev3", 00:14:34.628 "uuid": "060b1088-d8db-4c57-8fd5-5e0fc40f0ccd", 00:14:34.628 "is_configured": true, 00:14:34.628 "data_offset": 2048, 00:14:34.628 "data_size": 63488 00:14:34.628 }, 00:14:34.628 { 00:14:34.628 "name": "BaseBdev4", 00:14:34.628 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:34.628 "is_configured": false, 00:14:34.628 "data_offset": 0, 00:14:34.628 "data_size": 0 00:14:34.628 } 00:14:34.628 ] 00:14:34.628 }' 00:14:34.628 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:34.628 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:34.886 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:34.886 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:34.886 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.146 [2024-12-05 19:04:52.455312] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:35.146 BaseBdev4 00:14:35.146 [2024-12-05 19:04:52.455618] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:35.146 [2024-12-05 19:04:52.455653] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:35.147 [2024-12-05 19:04:52.455964] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:35.147 [2024-12-05 19:04:52.456454] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:35.147 [2024-12-05 19:04:52.456470] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:14:35.147 [2024-12-05 19:04:52.456598] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.147 [ 00:14:35.147 { 00:14:35.147 "name": "BaseBdev4", 00:14:35.147 "aliases": [ 00:14:35.147 "27888711-f5b2-40ed-8388-b44168d5d01b" 00:14:35.147 ], 00:14:35.147 "product_name": "Malloc disk", 00:14:35.147 "block_size": 512, 00:14:35.147 "num_blocks": 65536, 00:14:35.147 "uuid": "27888711-f5b2-40ed-8388-b44168d5d01b", 00:14:35.147 "assigned_rate_limits": { 00:14:35.147 "rw_ios_per_sec": 0, 00:14:35.147 "rw_mbytes_per_sec": 0, 00:14:35.147 "r_mbytes_per_sec": 0, 00:14:35.147 "w_mbytes_per_sec": 0 00:14:35.147 }, 00:14:35.147 "claimed": true, 00:14:35.147 "claim_type": "exclusive_write", 00:14:35.147 "zoned": false, 00:14:35.147 "supported_io_types": { 00:14:35.147 "read": true, 00:14:35.147 "write": true, 00:14:35.147 "unmap": true, 00:14:35.147 "flush": true, 00:14:35.147 "reset": true, 00:14:35.147 "nvme_admin": false, 00:14:35.147 "nvme_io": false, 00:14:35.147 "nvme_io_md": false, 00:14:35.147 "write_zeroes": true, 00:14:35.147 "zcopy": true, 00:14:35.147 "get_zone_info": false, 00:14:35.147 "zone_management": false, 00:14:35.147 "zone_append": false, 00:14:35.147 "compare": false, 00:14:35.147 "compare_and_write": false, 00:14:35.147 "abort": true, 00:14:35.147 "seek_hole": false, 00:14:35.147 "seek_data": false, 00:14:35.147 "copy": true, 00:14:35.147 "nvme_iov_md": false 00:14:35.147 }, 00:14:35.147 "memory_domains": [ 00:14:35.147 { 00:14:35.147 "dma_device_id": "system", 00:14:35.147 "dma_device_type": 1 00:14:35.147 }, 00:14:35.147 { 00:14:35.147 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:35.147 "dma_device_type": 2 00:14:35.147 } 00:14:35.147 ], 00:14:35.147 "driver_specific": {} 00:14:35.147 } 00:14:35.147 ] 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:35.147 "name": "Existed_Raid", 00:14:35.147 "uuid": "3bb29c02-0c83-4dd4-80f1-fe8cc0c49222", 00:14:35.147 "strip_size_kb": 64, 00:14:35.147 "state": "online", 00:14:35.147 "raid_level": "raid5f", 00:14:35.147 "superblock": true, 00:14:35.147 "num_base_bdevs": 4, 00:14:35.147 "num_base_bdevs_discovered": 4, 00:14:35.147 "num_base_bdevs_operational": 4, 00:14:35.147 "base_bdevs_list": [ 00:14:35.147 { 00:14:35.147 "name": "BaseBdev1", 00:14:35.147 "uuid": "663442b2-0227-4550-9aeb-431bea4b2e8c", 00:14:35.147 "is_configured": true, 00:14:35.147 "data_offset": 2048, 00:14:35.147 "data_size": 63488 00:14:35.147 }, 00:14:35.147 { 00:14:35.147 "name": "BaseBdev2", 00:14:35.147 "uuid": "122672ca-f7a4-4366-80ea-c7da78ee7ae5", 00:14:35.147 "is_configured": true, 00:14:35.147 "data_offset": 2048, 00:14:35.147 "data_size": 63488 00:14:35.147 }, 00:14:35.147 { 00:14:35.147 "name": "BaseBdev3", 00:14:35.147 "uuid": "060b1088-d8db-4c57-8fd5-5e0fc40f0ccd", 00:14:35.147 "is_configured": true, 00:14:35.147 "data_offset": 2048, 00:14:35.147 "data_size": 63488 00:14:35.147 }, 00:14:35.147 { 00:14:35.147 "name": "BaseBdev4", 00:14:35.147 "uuid": "27888711-f5b2-40ed-8388-b44168d5d01b", 00:14:35.147 "is_configured": true, 00:14:35.147 "data_offset": 2048, 00:14:35.147 "data_size": 63488 00:14:35.147 } 00:14:35.147 ] 00:14:35.147 }' 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:35.147 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.407 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:35.407 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:35.407 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:35.407 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:35.407 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:35.407 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:35.407 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:35.407 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:35.407 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.407 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.407 [2024-12-05 19:04:52.914744] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:35.407 19:04:52 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.407 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:35.407 "name": "Existed_Raid", 00:14:35.407 "aliases": [ 00:14:35.407 "3bb29c02-0c83-4dd4-80f1-fe8cc0c49222" 00:14:35.407 ], 00:14:35.407 "product_name": "Raid Volume", 00:14:35.407 "block_size": 512, 00:14:35.407 "num_blocks": 190464, 00:14:35.407 "uuid": "3bb29c02-0c83-4dd4-80f1-fe8cc0c49222", 00:14:35.407 "assigned_rate_limits": { 00:14:35.407 "rw_ios_per_sec": 0, 00:14:35.407 "rw_mbytes_per_sec": 0, 00:14:35.407 "r_mbytes_per_sec": 0, 00:14:35.407 "w_mbytes_per_sec": 0 00:14:35.407 }, 00:14:35.407 "claimed": false, 00:14:35.408 "zoned": false, 00:14:35.408 "supported_io_types": { 00:14:35.408 "read": true, 00:14:35.408 "write": true, 00:14:35.408 "unmap": false, 00:14:35.408 "flush": false, 00:14:35.408 "reset": true, 00:14:35.408 "nvme_admin": false, 00:14:35.408 "nvme_io": false, 00:14:35.408 "nvme_io_md": false, 00:14:35.408 "write_zeroes": true, 00:14:35.408 "zcopy": false, 00:14:35.408 "get_zone_info": false, 00:14:35.408 "zone_management": false, 00:14:35.408 "zone_append": false, 00:14:35.408 "compare": false, 00:14:35.408 "compare_and_write": false, 00:14:35.408 "abort": false, 00:14:35.408 "seek_hole": false, 00:14:35.408 "seek_data": false, 00:14:35.408 "copy": false, 00:14:35.408 "nvme_iov_md": false 00:14:35.408 }, 00:14:35.408 "driver_specific": { 00:14:35.408 "raid": { 00:14:35.408 "uuid": "3bb29c02-0c83-4dd4-80f1-fe8cc0c49222", 00:14:35.408 "strip_size_kb": 64, 00:14:35.408 "state": "online", 00:14:35.408 "raid_level": "raid5f", 00:14:35.408 "superblock": true, 00:14:35.408 "num_base_bdevs": 4, 00:14:35.408 "num_base_bdevs_discovered": 4, 00:14:35.408 "num_base_bdevs_operational": 4, 00:14:35.408 "base_bdevs_list": [ 00:14:35.408 { 00:14:35.408 "name": "BaseBdev1", 00:14:35.408 "uuid": "663442b2-0227-4550-9aeb-431bea4b2e8c", 00:14:35.408 "is_configured": true, 00:14:35.408 "data_offset": 2048, 00:14:35.408 "data_size": 63488 00:14:35.408 }, 00:14:35.408 { 00:14:35.408 "name": "BaseBdev2", 00:14:35.408 "uuid": "122672ca-f7a4-4366-80ea-c7da78ee7ae5", 00:14:35.408 "is_configured": true, 00:14:35.408 "data_offset": 2048, 00:14:35.408 "data_size": 63488 00:14:35.408 }, 00:14:35.408 { 00:14:35.408 "name": "BaseBdev3", 00:14:35.408 "uuid": "060b1088-d8db-4c57-8fd5-5e0fc40f0ccd", 00:14:35.408 "is_configured": true, 00:14:35.408 "data_offset": 2048, 00:14:35.408 "data_size": 63488 00:14:35.408 }, 00:14:35.408 { 00:14:35.408 "name": "BaseBdev4", 00:14:35.408 "uuid": "27888711-f5b2-40ed-8388-b44168d5d01b", 00:14:35.408 "is_configured": true, 00:14:35.408 "data_offset": 2048, 00:14:35.408 "data_size": 63488 00:14:35.408 } 00:14:35.408 ] 00:14:35.408 } 00:14:35.408 } 00:14:35.408 }' 00:14:35.408 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:35.669 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:35.669 BaseBdev2 00:14:35.669 BaseBdev3 00:14:35.669 BaseBdev4' 00:14:35.669 19:04:52 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.669 [2024-12-05 19:04:53.198112] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:35.669 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.930 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:35.930 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:35.930 "name": "Existed_Raid", 00:14:35.930 "uuid": "3bb29c02-0c83-4dd4-80f1-fe8cc0c49222", 00:14:35.930 "strip_size_kb": 64, 00:14:35.930 "state": "online", 00:14:35.930 "raid_level": "raid5f", 00:14:35.930 "superblock": true, 00:14:35.930 "num_base_bdevs": 4, 00:14:35.930 "num_base_bdevs_discovered": 3, 00:14:35.930 "num_base_bdevs_operational": 3, 00:14:35.930 "base_bdevs_list": [ 00:14:35.930 { 00:14:35.930 "name": null, 00:14:35.930 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:35.930 "is_configured": false, 00:14:35.930 "data_offset": 0, 00:14:35.930 "data_size": 63488 00:14:35.930 }, 00:14:35.930 { 00:14:35.930 "name": "BaseBdev2", 00:14:35.930 "uuid": "122672ca-f7a4-4366-80ea-c7da78ee7ae5", 00:14:35.930 "is_configured": true, 00:14:35.930 "data_offset": 2048, 00:14:35.930 "data_size": 63488 00:14:35.930 }, 00:14:35.930 { 00:14:35.930 "name": "BaseBdev3", 00:14:35.930 "uuid": "060b1088-d8db-4c57-8fd5-5e0fc40f0ccd", 00:14:35.930 "is_configured": true, 00:14:35.930 "data_offset": 2048, 00:14:35.930 "data_size": 63488 00:14:35.930 }, 00:14:35.930 { 00:14:35.930 "name": "BaseBdev4", 00:14:35.930 "uuid": "27888711-f5b2-40ed-8388-b44168d5d01b", 00:14:35.930 "is_configured": true, 00:14:35.930 "data_offset": 2048, 00:14:35.930 "data_size": 63488 00:14:35.930 } 00:14:35.930 ] 00:14:35.930 }' 00:14:35.930 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:35.930 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.189 [2024-12-05 19:04:53.708630] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:36.189 [2024-12-05 19:04:53.708830] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:36.189 [2024-12-05 19:04:53.719855] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.189 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.449 [2024-12-05 19:04:53.763805] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.449 [2024-12-05 19:04:53.822542] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:14:36.449 [2024-12-05 19:04:53.822659] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.449 BaseBdev2 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.449 [ 00:14:36.449 { 00:14:36.449 "name": "BaseBdev2", 00:14:36.449 "aliases": [ 00:14:36.449 "c6ca3f2c-278a-480c-a1b7-ed9ec33fdd38" 00:14:36.449 ], 00:14:36.449 "product_name": "Malloc disk", 00:14:36.449 "block_size": 512, 00:14:36.449 "num_blocks": 65536, 00:14:36.449 "uuid": "c6ca3f2c-278a-480c-a1b7-ed9ec33fdd38", 00:14:36.449 "assigned_rate_limits": { 00:14:36.449 "rw_ios_per_sec": 0, 00:14:36.449 "rw_mbytes_per_sec": 0, 00:14:36.449 "r_mbytes_per_sec": 0, 00:14:36.449 "w_mbytes_per_sec": 0 00:14:36.449 }, 00:14:36.449 "claimed": false, 00:14:36.449 "zoned": false, 00:14:36.449 "supported_io_types": { 00:14:36.449 "read": true, 00:14:36.449 "write": true, 00:14:36.449 "unmap": true, 00:14:36.449 "flush": true, 00:14:36.449 "reset": true, 00:14:36.449 "nvme_admin": false, 00:14:36.449 "nvme_io": false, 00:14:36.449 "nvme_io_md": false, 00:14:36.449 "write_zeroes": true, 00:14:36.449 "zcopy": true, 00:14:36.449 "get_zone_info": false, 00:14:36.449 "zone_management": false, 00:14:36.449 "zone_append": false, 00:14:36.449 "compare": false, 00:14:36.449 "compare_and_write": false, 00:14:36.449 "abort": true, 00:14:36.449 "seek_hole": false, 00:14:36.449 "seek_data": false, 00:14:36.449 "copy": true, 00:14:36.449 "nvme_iov_md": false 00:14:36.449 }, 00:14:36.449 "memory_domains": [ 00:14:36.449 { 00:14:36.449 "dma_device_id": "system", 00:14:36.449 "dma_device_type": 1 00:14:36.449 }, 00:14:36.449 { 00:14:36.449 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:36.449 "dma_device_type": 2 00:14:36.449 } 00:14:36.449 ], 00:14:36.449 "driver_specific": {} 00:14:36.449 } 00:14:36.449 ] 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.449 BaseBdev3 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev3 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.449 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.449 [ 00:14:36.449 { 00:14:36.449 "name": "BaseBdev3", 00:14:36.449 "aliases": [ 00:14:36.449 "44e1a930-0021-4dca-8c0d-c779b66500b0" 00:14:36.449 ], 00:14:36.449 "product_name": "Malloc disk", 00:14:36.449 "block_size": 512, 00:14:36.449 "num_blocks": 65536, 00:14:36.449 "uuid": "44e1a930-0021-4dca-8c0d-c779b66500b0", 00:14:36.449 "assigned_rate_limits": { 00:14:36.449 "rw_ios_per_sec": 0, 00:14:36.449 "rw_mbytes_per_sec": 0, 00:14:36.449 "r_mbytes_per_sec": 0, 00:14:36.449 "w_mbytes_per_sec": 0 00:14:36.449 }, 00:14:36.449 "claimed": false, 00:14:36.449 "zoned": false, 00:14:36.449 "supported_io_types": { 00:14:36.449 "read": true, 00:14:36.449 "write": true, 00:14:36.449 "unmap": true, 00:14:36.449 "flush": true, 00:14:36.449 "reset": true, 00:14:36.449 "nvme_admin": false, 00:14:36.450 "nvme_io": false, 00:14:36.450 "nvme_io_md": false, 00:14:36.450 "write_zeroes": true, 00:14:36.450 "zcopy": true, 00:14:36.450 "get_zone_info": false, 00:14:36.450 "zone_management": false, 00:14:36.450 "zone_append": false, 00:14:36.450 "compare": false, 00:14:36.450 "compare_and_write": false, 00:14:36.450 "abort": true, 00:14:36.450 "seek_hole": false, 00:14:36.450 "seek_data": false, 00:14:36.450 "copy": true, 00:14:36.450 "nvme_iov_md": false 00:14:36.450 }, 00:14:36.450 "memory_domains": [ 00:14:36.450 { 00:14:36.450 "dma_device_id": "system", 00:14:36.450 "dma_device_type": 1 00:14:36.450 }, 00:14:36.450 { 00:14:36.450 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:36.450 "dma_device_type": 2 00:14:36.450 } 00:14:36.450 ], 00:14:36.450 "driver_specific": {} 00:14:36.450 } 00:14:36.450 ] 00:14:36.450 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.450 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:36.450 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:36.450 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:36.450 19:04:53 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:36.450 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.450 19:04:53 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.450 BaseBdev4 00:14:36.450 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.450 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:14:36.450 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev4 00:14:36.450 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:36.450 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:36.450 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:36.450 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:36.450 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:36.450 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.450 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.710 [ 00:14:36.710 { 00:14:36.710 "name": "BaseBdev4", 00:14:36.710 "aliases": [ 00:14:36.710 "7c8aa143-2983-40a2-a37f-36651a0f844a" 00:14:36.710 ], 00:14:36.710 "product_name": "Malloc disk", 00:14:36.710 "block_size": 512, 00:14:36.710 "num_blocks": 65536, 00:14:36.710 "uuid": "7c8aa143-2983-40a2-a37f-36651a0f844a", 00:14:36.710 "assigned_rate_limits": { 00:14:36.710 "rw_ios_per_sec": 0, 00:14:36.710 "rw_mbytes_per_sec": 0, 00:14:36.710 "r_mbytes_per_sec": 0, 00:14:36.710 "w_mbytes_per_sec": 0 00:14:36.710 }, 00:14:36.710 "claimed": false, 00:14:36.710 "zoned": false, 00:14:36.710 "supported_io_types": { 00:14:36.710 "read": true, 00:14:36.710 "write": true, 00:14:36.710 "unmap": true, 00:14:36.710 "flush": true, 00:14:36.710 "reset": true, 00:14:36.710 "nvme_admin": false, 00:14:36.710 "nvme_io": false, 00:14:36.710 "nvme_io_md": false, 00:14:36.710 "write_zeroes": true, 00:14:36.710 "zcopy": true, 00:14:36.710 "get_zone_info": false, 00:14:36.710 "zone_management": false, 00:14:36.710 "zone_append": false, 00:14:36.710 "compare": false, 00:14:36.710 "compare_and_write": false, 00:14:36.710 "abort": true, 00:14:36.710 "seek_hole": false, 00:14:36.710 "seek_data": false, 00:14:36.710 "copy": true, 00:14:36.710 "nvme_iov_md": false 00:14:36.710 }, 00:14:36.710 "memory_domains": [ 00:14:36.710 { 00:14:36.710 "dma_device_id": "system", 00:14:36.710 "dma_device_type": 1 00:14:36.710 }, 00:14:36.710 { 00:14:36.710 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:36.710 "dma_device_type": 2 00:14:36.710 } 00:14:36.710 ], 00:14:36.710 "driver_specific": {} 00:14:36.710 } 00:14:36.710 ] 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.710 [2024-12-05 19:04:54.044947] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:36.710 [2024-12-05 19:04:54.045082] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:36.710 [2024-12-05 19:04:54.045145] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:36.710 [2024-12-05 19:04:54.046966] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:36.710 [2024-12-05 19:04:54.047064] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:36.710 "name": "Existed_Raid", 00:14:36.710 "uuid": "6e38e4e1-82b1-41fe-a0af-0ef773e78a7d", 00:14:36.710 "strip_size_kb": 64, 00:14:36.710 "state": "configuring", 00:14:36.710 "raid_level": "raid5f", 00:14:36.710 "superblock": true, 00:14:36.710 "num_base_bdevs": 4, 00:14:36.710 "num_base_bdevs_discovered": 3, 00:14:36.710 "num_base_bdevs_operational": 4, 00:14:36.710 "base_bdevs_list": [ 00:14:36.710 { 00:14:36.710 "name": "BaseBdev1", 00:14:36.710 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:36.710 "is_configured": false, 00:14:36.710 "data_offset": 0, 00:14:36.710 "data_size": 0 00:14:36.710 }, 00:14:36.710 { 00:14:36.710 "name": "BaseBdev2", 00:14:36.710 "uuid": "c6ca3f2c-278a-480c-a1b7-ed9ec33fdd38", 00:14:36.710 "is_configured": true, 00:14:36.710 "data_offset": 2048, 00:14:36.710 "data_size": 63488 00:14:36.710 }, 00:14:36.710 { 00:14:36.710 "name": "BaseBdev3", 00:14:36.710 "uuid": "44e1a930-0021-4dca-8c0d-c779b66500b0", 00:14:36.710 "is_configured": true, 00:14:36.710 "data_offset": 2048, 00:14:36.710 "data_size": 63488 00:14:36.710 }, 00:14:36.710 { 00:14:36.710 "name": "BaseBdev4", 00:14:36.710 "uuid": "7c8aa143-2983-40a2-a37f-36651a0f844a", 00:14:36.710 "is_configured": true, 00:14:36.710 "data_offset": 2048, 00:14:36.710 "data_size": 63488 00:14:36.710 } 00:14:36.710 ] 00:14:36.710 }' 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:36.710 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.969 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:36.969 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.969 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.969 [2024-12-05 19:04:54.508140] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:36.970 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:36.970 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:36.970 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:36.970 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:36.970 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:36.970 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:36.970 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:36.970 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:36.970 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:36.970 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:36.970 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:36.970 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.970 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:36.970 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:36.970 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.228 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.228 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:37.228 "name": "Existed_Raid", 00:14:37.228 "uuid": "6e38e4e1-82b1-41fe-a0af-0ef773e78a7d", 00:14:37.228 "strip_size_kb": 64, 00:14:37.228 "state": "configuring", 00:14:37.228 "raid_level": "raid5f", 00:14:37.228 "superblock": true, 00:14:37.228 "num_base_bdevs": 4, 00:14:37.228 "num_base_bdevs_discovered": 2, 00:14:37.228 "num_base_bdevs_operational": 4, 00:14:37.228 "base_bdevs_list": [ 00:14:37.228 { 00:14:37.228 "name": "BaseBdev1", 00:14:37.228 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:37.228 "is_configured": false, 00:14:37.228 "data_offset": 0, 00:14:37.228 "data_size": 0 00:14:37.228 }, 00:14:37.228 { 00:14:37.228 "name": null, 00:14:37.228 "uuid": "c6ca3f2c-278a-480c-a1b7-ed9ec33fdd38", 00:14:37.228 "is_configured": false, 00:14:37.228 "data_offset": 0, 00:14:37.228 "data_size": 63488 00:14:37.228 }, 00:14:37.228 { 00:14:37.228 "name": "BaseBdev3", 00:14:37.228 "uuid": "44e1a930-0021-4dca-8c0d-c779b66500b0", 00:14:37.228 "is_configured": true, 00:14:37.228 "data_offset": 2048, 00:14:37.228 "data_size": 63488 00:14:37.228 }, 00:14:37.228 { 00:14:37.228 "name": "BaseBdev4", 00:14:37.228 "uuid": "7c8aa143-2983-40a2-a37f-36651a0f844a", 00:14:37.228 "is_configured": true, 00:14:37.228 "data_offset": 2048, 00:14:37.228 "data_size": 63488 00:14:37.228 } 00:14:37.228 ] 00:14:37.228 }' 00:14:37.228 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:37.228 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.487 [2024-12-05 19:04:54.958272] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:37.487 BaseBdev1 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.487 [ 00:14:37.487 { 00:14:37.487 "name": "BaseBdev1", 00:14:37.487 "aliases": [ 00:14:37.487 "7e2a2890-aa65-4a1c-84ae-9e6ec92443e6" 00:14:37.487 ], 00:14:37.487 "product_name": "Malloc disk", 00:14:37.487 "block_size": 512, 00:14:37.487 "num_blocks": 65536, 00:14:37.487 "uuid": "7e2a2890-aa65-4a1c-84ae-9e6ec92443e6", 00:14:37.487 "assigned_rate_limits": { 00:14:37.487 "rw_ios_per_sec": 0, 00:14:37.487 "rw_mbytes_per_sec": 0, 00:14:37.487 "r_mbytes_per_sec": 0, 00:14:37.487 "w_mbytes_per_sec": 0 00:14:37.487 }, 00:14:37.487 "claimed": true, 00:14:37.487 "claim_type": "exclusive_write", 00:14:37.487 "zoned": false, 00:14:37.487 "supported_io_types": { 00:14:37.487 "read": true, 00:14:37.487 "write": true, 00:14:37.487 "unmap": true, 00:14:37.487 "flush": true, 00:14:37.487 "reset": true, 00:14:37.487 "nvme_admin": false, 00:14:37.487 "nvme_io": false, 00:14:37.487 "nvme_io_md": false, 00:14:37.487 "write_zeroes": true, 00:14:37.487 "zcopy": true, 00:14:37.487 "get_zone_info": false, 00:14:37.487 "zone_management": false, 00:14:37.487 "zone_append": false, 00:14:37.487 "compare": false, 00:14:37.487 "compare_and_write": false, 00:14:37.487 "abort": true, 00:14:37.487 "seek_hole": false, 00:14:37.487 "seek_data": false, 00:14:37.487 "copy": true, 00:14:37.487 "nvme_iov_md": false 00:14:37.487 }, 00:14:37.487 "memory_domains": [ 00:14:37.487 { 00:14:37.487 "dma_device_id": "system", 00:14:37.487 "dma_device_type": 1 00:14:37.487 }, 00:14:37.487 { 00:14:37.487 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:37.487 "dma_device_type": 2 00:14:37.487 } 00:14:37.487 ], 00:14:37.487 "driver_specific": {} 00:14:37.487 } 00:14:37.487 ] 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:37.487 19:04:54 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:37.487 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.487 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:37.487 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:37.487 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.487 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:37.746 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:37.746 "name": "Existed_Raid", 00:14:37.746 "uuid": "6e38e4e1-82b1-41fe-a0af-0ef773e78a7d", 00:14:37.746 "strip_size_kb": 64, 00:14:37.746 "state": "configuring", 00:14:37.746 "raid_level": "raid5f", 00:14:37.746 "superblock": true, 00:14:37.746 "num_base_bdevs": 4, 00:14:37.746 "num_base_bdevs_discovered": 3, 00:14:37.746 "num_base_bdevs_operational": 4, 00:14:37.746 "base_bdevs_list": [ 00:14:37.746 { 00:14:37.746 "name": "BaseBdev1", 00:14:37.746 "uuid": "7e2a2890-aa65-4a1c-84ae-9e6ec92443e6", 00:14:37.746 "is_configured": true, 00:14:37.746 "data_offset": 2048, 00:14:37.746 "data_size": 63488 00:14:37.746 }, 00:14:37.746 { 00:14:37.746 "name": null, 00:14:37.746 "uuid": "c6ca3f2c-278a-480c-a1b7-ed9ec33fdd38", 00:14:37.746 "is_configured": false, 00:14:37.746 "data_offset": 0, 00:14:37.746 "data_size": 63488 00:14:37.746 }, 00:14:37.746 { 00:14:37.746 "name": "BaseBdev3", 00:14:37.746 "uuid": "44e1a930-0021-4dca-8c0d-c779b66500b0", 00:14:37.746 "is_configured": true, 00:14:37.746 "data_offset": 2048, 00:14:37.746 "data_size": 63488 00:14:37.746 }, 00:14:37.746 { 00:14:37.746 "name": "BaseBdev4", 00:14:37.746 "uuid": "7c8aa143-2983-40a2-a37f-36651a0f844a", 00:14:37.746 "is_configured": true, 00:14:37.746 "data_offset": 2048, 00:14:37.746 "data_size": 63488 00:14:37.746 } 00:14:37.746 ] 00:14:37.746 }' 00:14:37.746 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:37.746 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:38.005 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.005 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:38.005 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.005 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:38.005 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.005 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:38.005 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:38.005 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.005 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:38.005 [2024-12-05 19:04:55.525343] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:38.005 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.005 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:38.006 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:38.006 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:38.006 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:38.006 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:38.006 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:38.006 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:38.006 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:38.006 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:38.006 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:38.006 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.006 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:38.006 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.006 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:38.006 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.264 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:38.264 "name": "Existed_Raid", 00:14:38.264 "uuid": "6e38e4e1-82b1-41fe-a0af-0ef773e78a7d", 00:14:38.264 "strip_size_kb": 64, 00:14:38.264 "state": "configuring", 00:14:38.264 "raid_level": "raid5f", 00:14:38.264 "superblock": true, 00:14:38.264 "num_base_bdevs": 4, 00:14:38.264 "num_base_bdevs_discovered": 2, 00:14:38.264 "num_base_bdevs_operational": 4, 00:14:38.264 "base_bdevs_list": [ 00:14:38.264 { 00:14:38.264 "name": "BaseBdev1", 00:14:38.264 "uuid": "7e2a2890-aa65-4a1c-84ae-9e6ec92443e6", 00:14:38.264 "is_configured": true, 00:14:38.264 "data_offset": 2048, 00:14:38.264 "data_size": 63488 00:14:38.264 }, 00:14:38.264 { 00:14:38.264 "name": null, 00:14:38.264 "uuid": "c6ca3f2c-278a-480c-a1b7-ed9ec33fdd38", 00:14:38.264 "is_configured": false, 00:14:38.264 "data_offset": 0, 00:14:38.264 "data_size": 63488 00:14:38.264 }, 00:14:38.264 { 00:14:38.264 "name": null, 00:14:38.264 "uuid": "44e1a930-0021-4dca-8c0d-c779b66500b0", 00:14:38.264 "is_configured": false, 00:14:38.264 "data_offset": 0, 00:14:38.264 "data_size": 63488 00:14:38.264 }, 00:14:38.264 { 00:14:38.264 "name": "BaseBdev4", 00:14:38.264 "uuid": "7c8aa143-2983-40a2-a37f-36651a0f844a", 00:14:38.264 "is_configured": true, 00:14:38.264 "data_offset": 2048, 00:14:38.264 "data_size": 63488 00:14:38.264 } 00:14:38.264 ] 00:14:38.264 }' 00:14:38.264 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:38.264 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:38.522 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.522 19:04:55 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:38.522 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.522 19:04:55 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:38.522 [2024-12-05 19:04:56.032490] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:38.522 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:38.781 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:38.781 "name": "Existed_Raid", 00:14:38.781 "uuid": "6e38e4e1-82b1-41fe-a0af-0ef773e78a7d", 00:14:38.781 "strip_size_kb": 64, 00:14:38.781 "state": "configuring", 00:14:38.781 "raid_level": "raid5f", 00:14:38.781 "superblock": true, 00:14:38.781 "num_base_bdevs": 4, 00:14:38.781 "num_base_bdevs_discovered": 3, 00:14:38.781 "num_base_bdevs_operational": 4, 00:14:38.781 "base_bdevs_list": [ 00:14:38.781 { 00:14:38.781 "name": "BaseBdev1", 00:14:38.781 "uuid": "7e2a2890-aa65-4a1c-84ae-9e6ec92443e6", 00:14:38.781 "is_configured": true, 00:14:38.781 "data_offset": 2048, 00:14:38.781 "data_size": 63488 00:14:38.781 }, 00:14:38.781 { 00:14:38.781 "name": null, 00:14:38.781 "uuid": "c6ca3f2c-278a-480c-a1b7-ed9ec33fdd38", 00:14:38.781 "is_configured": false, 00:14:38.781 "data_offset": 0, 00:14:38.781 "data_size": 63488 00:14:38.781 }, 00:14:38.781 { 00:14:38.781 "name": "BaseBdev3", 00:14:38.781 "uuid": "44e1a930-0021-4dca-8c0d-c779b66500b0", 00:14:38.781 "is_configured": true, 00:14:38.781 "data_offset": 2048, 00:14:38.781 "data_size": 63488 00:14:38.781 }, 00:14:38.781 { 00:14:38.781 "name": "BaseBdev4", 00:14:38.781 "uuid": "7c8aa143-2983-40a2-a37f-36651a0f844a", 00:14:38.781 "is_configured": true, 00:14:38.781 "data_offset": 2048, 00:14:38.781 "data_size": 63488 00:14:38.781 } 00:14:38.781 ] 00:14:38.781 }' 00:14:38.781 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:38.781 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.038 [2024-12-05 19:04:56.531827] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:39.038 "name": "Existed_Raid", 00:14:39.038 "uuid": "6e38e4e1-82b1-41fe-a0af-0ef773e78a7d", 00:14:39.038 "strip_size_kb": 64, 00:14:39.038 "state": "configuring", 00:14:39.038 "raid_level": "raid5f", 00:14:39.038 "superblock": true, 00:14:39.038 "num_base_bdevs": 4, 00:14:39.038 "num_base_bdevs_discovered": 2, 00:14:39.038 "num_base_bdevs_operational": 4, 00:14:39.038 "base_bdevs_list": [ 00:14:39.038 { 00:14:39.038 "name": null, 00:14:39.038 "uuid": "7e2a2890-aa65-4a1c-84ae-9e6ec92443e6", 00:14:39.038 "is_configured": false, 00:14:39.038 "data_offset": 0, 00:14:39.038 "data_size": 63488 00:14:39.038 }, 00:14:39.038 { 00:14:39.038 "name": null, 00:14:39.038 "uuid": "c6ca3f2c-278a-480c-a1b7-ed9ec33fdd38", 00:14:39.038 "is_configured": false, 00:14:39.038 "data_offset": 0, 00:14:39.038 "data_size": 63488 00:14:39.038 }, 00:14:39.038 { 00:14:39.038 "name": "BaseBdev3", 00:14:39.038 "uuid": "44e1a930-0021-4dca-8c0d-c779b66500b0", 00:14:39.038 "is_configured": true, 00:14:39.038 "data_offset": 2048, 00:14:39.038 "data_size": 63488 00:14:39.038 }, 00:14:39.038 { 00:14:39.038 "name": "BaseBdev4", 00:14:39.038 "uuid": "7c8aa143-2983-40a2-a37f-36651a0f844a", 00:14:39.038 "is_configured": true, 00:14:39.038 "data_offset": 2048, 00:14:39.038 "data_size": 63488 00:14:39.038 } 00:14:39.038 ] 00:14:39.038 }' 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:39.038 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.605 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.605 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:39.605 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.605 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.605 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.605 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:39.605 19:04:56 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:39.605 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.605 19:04:56 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.605 [2024-12-05 19:04:57.005489] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:39.605 "name": "Existed_Raid", 00:14:39.605 "uuid": "6e38e4e1-82b1-41fe-a0af-0ef773e78a7d", 00:14:39.605 "strip_size_kb": 64, 00:14:39.605 "state": "configuring", 00:14:39.605 "raid_level": "raid5f", 00:14:39.605 "superblock": true, 00:14:39.605 "num_base_bdevs": 4, 00:14:39.605 "num_base_bdevs_discovered": 3, 00:14:39.605 "num_base_bdevs_operational": 4, 00:14:39.605 "base_bdevs_list": [ 00:14:39.605 { 00:14:39.605 "name": null, 00:14:39.605 "uuid": "7e2a2890-aa65-4a1c-84ae-9e6ec92443e6", 00:14:39.605 "is_configured": false, 00:14:39.605 "data_offset": 0, 00:14:39.605 "data_size": 63488 00:14:39.605 }, 00:14:39.605 { 00:14:39.605 "name": "BaseBdev2", 00:14:39.605 "uuid": "c6ca3f2c-278a-480c-a1b7-ed9ec33fdd38", 00:14:39.605 "is_configured": true, 00:14:39.605 "data_offset": 2048, 00:14:39.605 "data_size": 63488 00:14:39.605 }, 00:14:39.605 { 00:14:39.605 "name": "BaseBdev3", 00:14:39.605 "uuid": "44e1a930-0021-4dca-8c0d-c779b66500b0", 00:14:39.605 "is_configured": true, 00:14:39.605 "data_offset": 2048, 00:14:39.605 "data_size": 63488 00:14:39.605 }, 00:14:39.605 { 00:14:39.605 "name": "BaseBdev4", 00:14:39.605 "uuid": "7c8aa143-2983-40a2-a37f-36651a0f844a", 00:14:39.605 "is_configured": true, 00:14:39.605 "data_offset": 2048, 00:14:39.605 "data_size": 63488 00:14:39.605 } 00:14:39.605 ] 00:14:39.605 }' 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:39.605 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 7e2a2890-aa65-4a1c-84ae-9e6ec92443e6 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.176 [2024-12-05 19:04:57.606548] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:40.176 [2024-12-05 19:04:57.606806] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:40.176 [2024-12-05 19:04:57.606856] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:40.176 [2024-12-05 19:04:57.607130] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:14:40.176 NewBaseBdev 00:14:40.176 [2024-12-05 19:04:57.607599] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:40.176 [2024-12-05 19:04:57.607621] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:14:40.176 [2024-12-05 19:04:57.607731] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@903 -- # local bdev_name=NewBaseBdev 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@905 -- # local i 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.176 [ 00:14:40.176 { 00:14:40.176 "name": "NewBaseBdev", 00:14:40.176 "aliases": [ 00:14:40.176 "7e2a2890-aa65-4a1c-84ae-9e6ec92443e6" 00:14:40.176 ], 00:14:40.176 "product_name": "Malloc disk", 00:14:40.176 "block_size": 512, 00:14:40.176 "num_blocks": 65536, 00:14:40.176 "uuid": "7e2a2890-aa65-4a1c-84ae-9e6ec92443e6", 00:14:40.176 "assigned_rate_limits": { 00:14:40.176 "rw_ios_per_sec": 0, 00:14:40.176 "rw_mbytes_per_sec": 0, 00:14:40.176 "r_mbytes_per_sec": 0, 00:14:40.176 "w_mbytes_per_sec": 0 00:14:40.176 }, 00:14:40.176 "claimed": true, 00:14:40.176 "claim_type": "exclusive_write", 00:14:40.176 "zoned": false, 00:14:40.176 "supported_io_types": { 00:14:40.176 "read": true, 00:14:40.176 "write": true, 00:14:40.176 "unmap": true, 00:14:40.176 "flush": true, 00:14:40.176 "reset": true, 00:14:40.176 "nvme_admin": false, 00:14:40.176 "nvme_io": false, 00:14:40.176 "nvme_io_md": false, 00:14:40.176 "write_zeroes": true, 00:14:40.176 "zcopy": true, 00:14:40.176 "get_zone_info": false, 00:14:40.176 "zone_management": false, 00:14:40.176 "zone_append": false, 00:14:40.176 "compare": false, 00:14:40.176 "compare_and_write": false, 00:14:40.176 "abort": true, 00:14:40.176 "seek_hole": false, 00:14:40.176 "seek_data": false, 00:14:40.176 "copy": true, 00:14:40.176 "nvme_iov_md": false 00:14:40.176 }, 00:14:40.176 "memory_domains": [ 00:14:40.176 { 00:14:40.176 "dma_device_id": "system", 00:14:40.176 "dma_device_type": 1 00:14:40.176 }, 00:14:40.176 { 00:14:40.176 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:40.176 "dma_device_type": 2 00:14:40.176 } 00:14:40.176 ], 00:14:40.176 "driver_specific": {} 00:14:40.176 } 00:14:40.176 ] 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@911 -- # return 0 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:40.176 "name": "Existed_Raid", 00:14:40.176 "uuid": "6e38e4e1-82b1-41fe-a0af-0ef773e78a7d", 00:14:40.176 "strip_size_kb": 64, 00:14:40.176 "state": "online", 00:14:40.176 "raid_level": "raid5f", 00:14:40.176 "superblock": true, 00:14:40.176 "num_base_bdevs": 4, 00:14:40.176 "num_base_bdevs_discovered": 4, 00:14:40.176 "num_base_bdevs_operational": 4, 00:14:40.176 "base_bdevs_list": [ 00:14:40.176 { 00:14:40.176 "name": "NewBaseBdev", 00:14:40.176 "uuid": "7e2a2890-aa65-4a1c-84ae-9e6ec92443e6", 00:14:40.176 "is_configured": true, 00:14:40.176 "data_offset": 2048, 00:14:40.176 "data_size": 63488 00:14:40.176 }, 00:14:40.176 { 00:14:40.176 "name": "BaseBdev2", 00:14:40.176 "uuid": "c6ca3f2c-278a-480c-a1b7-ed9ec33fdd38", 00:14:40.176 "is_configured": true, 00:14:40.176 "data_offset": 2048, 00:14:40.176 "data_size": 63488 00:14:40.176 }, 00:14:40.176 { 00:14:40.176 "name": "BaseBdev3", 00:14:40.176 "uuid": "44e1a930-0021-4dca-8c0d-c779b66500b0", 00:14:40.176 "is_configured": true, 00:14:40.176 "data_offset": 2048, 00:14:40.176 "data_size": 63488 00:14:40.176 }, 00:14:40.176 { 00:14:40.176 "name": "BaseBdev4", 00:14:40.176 "uuid": "7c8aa143-2983-40a2-a37f-36651a0f844a", 00:14:40.176 "is_configured": true, 00:14:40.176 "data_offset": 2048, 00:14:40.176 "data_size": 63488 00:14:40.176 } 00:14:40.176 ] 00:14:40.176 }' 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:40.176 19:04:57 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.747 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:40.747 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:40.747 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:40.747 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:40.747 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:40.747 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:40.747 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:40.747 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:40.747 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.747 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.747 [2024-12-05 19:04:58.081917] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:40.747 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.747 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:40.747 "name": "Existed_Raid", 00:14:40.747 "aliases": [ 00:14:40.747 "6e38e4e1-82b1-41fe-a0af-0ef773e78a7d" 00:14:40.747 ], 00:14:40.747 "product_name": "Raid Volume", 00:14:40.747 "block_size": 512, 00:14:40.747 "num_blocks": 190464, 00:14:40.747 "uuid": "6e38e4e1-82b1-41fe-a0af-0ef773e78a7d", 00:14:40.747 "assigned_rate_limits": { 00:14:40.747 "rw_ios_per_sec": 0, 00:14:40.747 "rw_mbytes_per_sec": 0, 00:14:40.747 "r_mbytes_per_sec": 0, 00:14:40.747 "w_mbytes_per_sec": 0 00:14:40.747 }, 00:14:40.747 "claimed": false, 00:14:40.747 "zoned": false, 00:14:40.747 "supported_io_types": { 00:14:40.747 "read": true, 00:14:40.747 "write": true, 00:14:40.747 "unmap": false, 00:14:40.747 "flush": false, 00:14:40.747 "reset": true, 00:14:40.747 "nvme_admin": false, 00:14:40.747 "nvme_io": false, 00:14:40.747 "nvme_io_md": false, 00:14:40.747 "write_zeroes": true, 00:14:40.747 "zcopy": false, 00:14:40.747 "get_zone_info": false, 00:14:40.747 "zone_management": false, 00:14:40.747 "zone_append": false, 00:14:40.747 "compare": false, 00:14:40.747 "compare_and_write": false, 00:14:40.747 "abort": false, 00:14:40.747 "seek_hole": false, 00:14:40.747 "seek_data": false, 00:14:40.747 "copy": false, 00:14:40.747 "nvme_iov_md": false 00:14:40.747 }, 00:14:40.747 "driver_specific": { 00:14:40.747 "raid": { 00:14:40.747 "uuid": "6e38e4e1-82b1-41fe-a0af-0ef773e78a7d", 00:14:40.747 "strip_size_kb": 64, 00:14:40.747 "state": "online", 00:14:40.747 "raid_level": "raid5f", 00:14:40.747 "superblock": true, 00:14:40.747 "num_base_bdevs": 4, 00:14:40.747 "num_base_bdevs_discovered": 4, 00:14:40.747 "num_base_bdevs_operational": 4, 00:14:40.747 "base_bdevs_list": [ 00:14:40.747 { 00:14:40.747 "name": "NewBaseBdev", 00:14:40.747 "uuid": "7e2a2890-aa65-4a1c-84ae-9e6ec92443e6", 00:14:40.747 "is_configured": true, 00:14:40.747 "data_offset": 2048, 00:14:40.747 "data_size": 63488 00:14:40.747 }, 00:14:40.747 { 00:14:40.747 "name": "BaseBdev2", 00:14:40.748 "uuid": "c6ca3f2c-278a-480c-a1b7-ed9ec33fdd38", 00:14:40.748 "is_configured": true, 00:14:40.748 "data_offset": 2048, 00:14:40.748 "data_size": 63488 00:14:40.748 }, 00:14:40.748 { 00:14:40.748 "name": "BaseBdev3", 00:14:40.748 "uuid": "44e1a930-0021-4dca-8c0d-c779b66500b0", 00:14:40.748 "is_configured": true, 00:14:40.748 "data_offset": 2048, 00:14:40.748 "data_size": 63488 00:14:40.748 }, 00:14:40.748 { 00:14:40.748 "name": "BaseBdev4", 00:14:40.748 "uuid": "7c8aa143-2983-40a2-a37f-36651a0f844a", 00:14:40.748 "is_configured": true, 00:14:40.748 "data_offset": 2048, 00:14:40.748 "data_size": 63488 00:14:40.748 } 00:14:40.748 ] 00:14:40.748 } 00:14:40.748 } 00:14:40.748 }' 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:40.748 BaseBdev2 00:14:40.748 BaseBdev3 00:14:40.748 BaseBdev4' 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:40.748 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.008 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.008 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:41.008 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:41.008 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:41.008 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:41.008 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:41.008 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.008 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.008 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.008 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:41.008 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:41.008 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:41.008 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.008 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.009 [2024-12-05 19:04:58.381290] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:41.009 [2024-12-05 19:04:58.381362] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:41.009 [2024-12-05 19:04:58.381458] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:41.009 [2024-12-05 19:04:58.381730] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:41.009 [2024-12-05 19:04:58.381788] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:14:41.009 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.009 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 91449 00:14:41.009 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # '[' -z 91449 ']' 00:14:41.009 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@958 -- # kill -0 91449 00:14:41.009 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # uname 00:14:41.009 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:41.009 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 91449 00:14:41.009 killing process with pid 91449 00:14:41.009 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:41.009 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:41.009 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 91449' 00:14:41.009 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@973 -- # kill 91449 00:14:41.009 [2024-12-05 19:04:58.429405] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:41.009 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@978 -- # wait 91449 00:14:41.009 [2024-12-05 19:04:58.469153] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:41.269 19:04:58 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:14:41.269 00:14:41.269 real 0m9.557s 00:14:41.269 user 0m16.348s 00:14:41.269 sys 0m2.097s 00:14:41.269 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:41.269 ************************************ 00:14:41.269 END TEST raid5f_state_function_test_sb 00:14:41.269 ************************************ 00:14:41.269 19:04:58 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.269 19:04:58 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 4 00:14:41.269 19:04:58 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:14:41.269 19:04:58 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:41.269 19:04:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:41.269 ************************************ 00:14:41.269 START TEST raid5f_superblock_test 00:14:41.269 ************************************ 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1129 -- # raid_superblock_test raid5f 4 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=92042 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 92042 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # '[' -z 92042 ']' 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:41.269 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:41.269 19:04:58 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:41.529 [2024-12-05 19:04:58.853703] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:14:41.529 [2024-12-05 19:04:58.853920] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92042 ] 00:14:41.529 [2024-12-05 19:04:59.009795] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:41.529 [2024-12-05 19:04:59.034807] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:41.529 [2024-12-05 19:04:59.078295] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:41.529 [2024-12-05 19:04:59.078332] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@868 -- # return 0 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.468 malloc1 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.468 [2024-12-05 19:04:59.694167] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:42.468 [2024-12-05 19:04:59.694316] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:42.468 [2024-12-05 19:04:59.694368] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:42.468 [2024-12-05 19:04:59.694402] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:42.468 [2024-12-05 19:04:59.696487] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:42.468 [2024-12-05 19:04:59.696567] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:42.468 pt1 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.468 malloc2 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.468 [2024-12-05 19:04:59.726724] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:42.468 [2024-12-05 19:04:59.726838] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:42.468 [2024-12-05 19:04:59.726889] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:42.468 [2024-12-05 19:04:59.726918] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:42.468 [2024-12-05 19:04:59.728948] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:42.468 [2024-12-05 19:04:59.729015] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:42.468 pt2 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:42.468 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.469 malloc3 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.469 [2024-12-05 19:04:59.759213] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:42.469 [2024-12-05 19:04:59.759322] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:42.469 [2024-12-05 19:04:59.759373] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:42.469 [2024-12-05 19:04:59.759401] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:42.469 [2024-12-05 19:04:59.761422] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:42.469 [2024-12-05 19:04:59.761494] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:42.469 pt3 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.469 malloc4 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.469 [2024-12-05 19:04:59.802797] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:42.469 [2024-12-05 19:04:59.803000] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:42.469 [2024-12-05 19:04:59.803076] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:42.469 [2024-12-05 19:04:59.803158] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:42.469 [2024-12-05 19:04:59.806779] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:42.469 [2024-12-05 19:04:59.806883] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:42.469 pt4 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.469 [2024-12-05 19:04:59.815102] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:42.469 [2024-12-05 19:04:59.817192] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:42.469 [2024-12-05 19:04:59.817310] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:42.469 [2024-12-05 19:04:59.817408] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:42.469 [2024-12-05 19:04:59.817633] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:42.469 [2024-12-05 19:04:59.817723] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:42.469 [2024-12-05 19:04:59.818022] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:42.469 [2024-12-05 19:04:59.818546] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:42.469 [2024-12-05 19:04:59.818602] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:42.469 [2024-12-05 19:04:59.818778] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:42.469 "name": "raid_bdev1", 00:14:42.469 "uuid": "0af56e1f-2d1a-4451-9269-e054d1ec2256", 00:14:42.469 "strip_size_kb": 64, 00:14:42.469 "state": "online", 00:14:42.469 "raid_level": "raid5f", 00:14:42.469 "superblock": true, 00:14:42.469 "num_base_bdevs": 4, 00:14:42.469 "num_base_bdevs_discovered": 4, 00:14:42.469 "num_base_bdevs_operational": 4, 00:14:42.469 "base_bdevs_list": [ 00:14:42.469 { 00:14:42.469 "name": "pt1", 00:14:42.469 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:42.469 "is_configured": true, 00:14:42.469 "data_offset": 2048, 00:14:42.469 "data_size": 63488 00:14:42.469 }, 00:14:42.469 { 00:14:42.469 "name": "pt2", 00:14:42.469 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:42.469 "is_configured": true, 00:14:42.469 "data_offset": 2048, 00:14:42.469 "data_size": 63488 00:14:42.469 }, 00:14:42.469 { 00:14:42.469 "name": "pt3", 00:14:42.469 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:42.469 "is_configured": true, 00:14:42.469 "data_offset": 2048, 00:14:42.469 "data_size": 63488 00:14:42.469 }, 00:14:42.469 { 00:14:42.469 "name": "pt4", 00:14:42.469 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:42.469 "is_configured": true, 00:14:42.469 "data_offset": 2048, 00:14:42.469 "data_size": 63488 00:14:42.469 } 00:14:42.469 ] 00:14:42.469 }' 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:42.469 19:04:59 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.729 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:14:42.729 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:42.729 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:42.729 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:42.729 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:42.729 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:42.729 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:42.729 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.729 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.729 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:42.990 [2024-12-05 19:05:00.288142] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:42.990 "name": "raid_bdev1", 00:14:42.990 "aliases": [ 00:14:42.990 "0af56e1f-2d1a-4451-9269-e054d1ec2256" 00:14:42.990 ], 00:14:42.990 "product_name": "Raid Volume", 00:14:42.990 "block_size": 512, 00:14:42.990 "num_blocks": 190464, 00:14:42.990 "uuid": "0af56e1f-2d1a-4451-9269-e054d1ec2256", 00:14:42.990 "assigned_rate_limits": { 00:14:42.990 "rw_ios_per_sec": 0, 00:14:42.990 "rw_mbytes_per_sec": 0, 00:14:42.990 "r_mbytes_per_sec": 0, 00:14:42.990 "w_mbytes_per_sec": 0 00:14:42.990 }, 00:14:42.990 "claimed": false, 00:14:42.990 "zoned": false, 00:14:42.990 "supported_io_types": { 00:14:42.990 "read": true, 00:14:42.990 "write": true, 00:14:42.990 "unmap": false, 00:14:42.990 "flush": false, 00:14:42.990 "reset": true, 00:14:42.990 "nvme_admin": false, 00:14:42.990 "nvme_io": false, 00:14:42.990 "nvme_io_md": false, 00:14:42.990 "write_zeroes": true, 00:14:42.990 "zcopy": false, 00:14:42.990 "get_zone_info": false, 00:14:42.990 "zone_management": false, 00:14:42.990 "zone_append": false, 00:14:42.990 "compare": false, 00:14:42.990 "compare_and_write": false, 00:14:42.990 "abort": false, 00:14:42.990 "seek_hole": false, 00:14:42.990 "seek_data": false, 00:14:42.990 "copy": false, 00:14:42.990 "nvme_iov_md": false 00:14:42.990 }, 00:14:42.990 "driver_specific": { 00:14:42.990 "raid": { 00:14:42.990 "uuid": "0af56e1f-2d1a-4451-9269-e054d1ec2256", 00:14:42.990 "strip_size_kb": 64, 00:14:42.990 "state": "online", 00:14:42.990 "raid_level": "raid5f", 00:14:42.990 "superblock": true, 00:14:42.990 "num_base_bdevs": 4, 00:14:42.990 "num_base_bdevs_discovered": 4, 00:14:42.990 "num_base_bdevs_operational": 4, 00:14:42.990 "base_bdevs_list": [ 00:14:42.990 { 00:14:42.990 "name": "pt1", 00:14:42.990 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:42.990 "is_configured": true, 00:14:42.990 "data_offset": 2048, 00:14:42.990 "data_size": 63488 00:14:42.990 }, 00:14:42.990 { 00:14:42.990 "name": "pt2", 00:14:42.990 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:42.990 "is_configured": true, 00:14:42.990 "data_offset": 2048, 00:14:42.990 "data_size": 63488 00:14:42.990 }, 00:14:42.990 { 00:14:42.990 "name": "pt3", 00:14:42.990 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:42.990 "is_configured": true, 00:14:42.990 "data_offset": 2048, 00:14:42.990 "data_size": 63488 00:14:42.990 }, 00:14:42.990 { 00:14:42.990 "name": "pt4", 00:14:42.990 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:42.990 "is_configured": true, 00:14:42.990 "data_offset": 2048, 00:14:42.990 "data_size": 63488 00:14:42.990 } 00:14:42.990 ] 00:14:42.990 } 00:14:42.990 } 00:14:42.990 }' 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:42.990 pt2 00:14:42.990 pt3 00:14:42.990 pt4' 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:42.990 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.250 [2024-12-05 19:05:00.607537] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=0af56e1f-2d1a-4451-9269-e054d1ec2256 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 0af56e1f-2d1a-4451-9269-e054d1ec2256 ']' 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.250 [2024-12-05 19:05:00.651316] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:43.250 [2024-12-05 19:05:00.651392] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:43.250 [2024-12-05 19:05:00.651463] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:43.250 [2024-12-05 19:05:00.651562] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:43.250 [2024-12-05 19:05:00.651593] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.250 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # local es=0 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:43.251 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.512 [2024-12-05 19:05:00.815049] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:14:43.512 [2024-12-05 19:05:00.816888] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:14:43.512 [2024-12-05 19:05:00.816970] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:14:43.512 [2024-12-05 19:05:00.817012] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:14:43.512 [2024-12-05 19:05:00.817086] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:14:43.512 [2024-12-05 19:05:00.817163] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:14:43.512 [2024-12-05 19:05:00.817208] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:14:43.512 [2024-12-05 19:05:00.817246] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:14:43.512 [2024-12-05 19:05:00.817287] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:43.512 [2024-12-05 19:05:00.817350] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:14:43.512 request: 00:14:43.512 { 00:14:43.512 "name": "raid_bdev1", 00:14:43.512 "raid_level": "raid5f", 00:14:43.512 "base_bdevs": [ 00:14:43.512 "malloc1", 00:14:43.512 "malloc2", 00:14:43.512 "malloc3", 00:14:43.512 "malloc4" 00:14:43.512 ], 00:14:43.512 "strip_size_kb": 64, 00:14:43.512 "superblock": false, 00:14:43.512 "method": "bdev_raid_create", 00:14:43.512 "req_id": 1 00:14:43.512 } 00:14:43.512 Got JSON-RPC error response 00:14:43.512 response: 00:14:43.512 { 00:14:43.512 "code": -17, 00:14:43.512 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:14:43.512 } 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@655 -- # es=1 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.512 [2024-12-05 19:05:00.882910] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:43.512 [2024-12-05 19:05:00.883005] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:43.512 [2024-12-05 19:05:00.883057] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:14:43.512 [2024-12-05 19:05:00.883084] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:43.512 [2024-12-05 19:05:00.885100] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:43.512 [2024-12-05 19:05:00.885164] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:43.512 [2024-12-05 19:05:00.885254] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:43.512 [2024-12-05 19:05:00.885310] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:43.512 pt1 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.512 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:43.512 "name": "raid_bdev1", 00:14:43.512 "uuid": "0af56e1f-2d1a-4451-9269-e054d1ec2256", 00:14:43.512 "strip_size_kb": 64, 00:14:43.512 "state": "configuring", 00:14:43.512 "raid_level": "raid5f", 00:14:43.512 "superblock": true, 00:14:43.512 "num_base_bdevs": 4, 00:14:43.512 "num_base_bdevs_discovered": 1, 00:14:43.512 "num_base_bdevs_operational": 4, 00:14:43.512 "base_bdevs_list": [ 00:14:43.512 { 00:14:43.512 "name": "pt1", 00:14:43.513 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:43.513 "is_configured": true, 00:14:43.513 "data_offset": 2048, 00:14:43.513 "data_size": 63488 00:14:43.513 }, 00:14:43.513 { 00:14:43.513 "name": null, 00:14:43.513 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:43.513 "is_configured": false, 00:14:43.513 "data_offset": 2048, 00:14:43.513 "data_size": 63488 00:14:43.513 }, 00:14:43.513 { 00:14:43.513 "name": null, 00:14:43.513 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:43.513 "is_configured": false, 00:14:43.513 "data_offset": 2048, 00:14:43.513 "data_size": 63488 00:14:43.513 }, 00:14:43.513 { 00:14:43.513 "name": null, 00:14:43.513 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:43.513 "is_configured": false, 00:14:43.513 "data_offset": 2048, 00:14:43.513 "data_size": 63488 00:14:43.513 } 00:14:43.513 ] 00:14:43.513 }' 00:14:43.513 19:05:00 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:43.513 19:05:00 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.774 [2024-12-05 19:05:01.306192] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:43.774 [2024-12-05 19:05:01.306284] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:43.774 [2024-12-05 19:05:01.306331] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:14:43.774 [2024-12-05 19:05:01.306357] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:43.774 [2024-12-05 19:05:01.306686] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:43.774 [2024-12-05 19:05:01.306739] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:43.774 [2024-12-05 19:05:01.306813] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:43.774 [2024-12-05 19:05:01.306859] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:43.774 pt2 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.774 [2024-12-05 19:05:01.318199] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:43.774 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.033 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.033 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.033 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:44.033 "name": "raid_bdev1", 00:14:44.033 "uuid": "0af56e1f-2d1a-4451-9269-e054d1ec2256", 00:14:44.033 "strip_size_kb": 64, 00:14:44.033 "state": "configuring", 00:14:44.033 "raid_level": "raid5f", 00:14:44.033 "superblock": true, 00:14:44.033 "num_base_bdevs": 4, 00:14:44.033 "num_base_bdevs_discovered": 1, 00:14:44.033 "num_base_bdevs_operational": 4, 00:14:44.033 "base_bdevs_list": [ 00:14:44.033 { 00:14:44.033 "name": "pt1", 00:14:44.033 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:44.033 "is_configured": true, 00:14:44.033 "data_offset": 2048, 00:14:44.033 "data_size": 63488 00:14:44.033 }, 00:14:44.033 { 00:14:44.033 "name": null, 00:14:44.033 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:44.033 "is_configured": false, 00:14:44.033 "data_offset": 0, 00:14:44.033 "data_size": 63488 00:14:44.033 }, 00:14:44.033 { 00:14:44.033 "name": null, 00:14:44.033 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:44.033 "is_configured": false, 00:14:44.033 "data_offset": 2048, 00:14:44.033 "data_size": 63488 00:14:44.033 }, 00:14:44.033 { 00:14:44.033 "name": null, 00:14:44.033 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:44.033 "is_configured": false, 00:14:44.033 "data_offset": 2048, 00:14:44.033 "data_size": 63488 00:14:44.033 } 00:14:44.033 ] 00:14:44.033 }' 00:14:44.033 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:44.033 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.293 [2024-12-05 19:05:01.741450] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:44.293 [2024-12-05 19:05:01.741547] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:44.293 [2024-12-05 19:05:01.741590] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:14:44.293 [2024-12-05 19:05:01.741619] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:44.293 [2024-12-05 19:05:01.741928] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:44.293 [2024-12-05 19:05:01.741986] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:44.293 [2024-12-05 19:05:01.742060] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:44.293 [2024-12-05 19:05:01.742115] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:44.293 pt2 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.293 [2024-12-05 19:05:01.753410] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:44.293 [2024-12-05 19:05:01.753456] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:44.293 [2024-12-05 19:05:01.753468] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:14:44.293 [2024-12-05 19:05:01.753477] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:44.293 [2024-12-05 19:05:01.753779] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:44.293 [2024-12-05 19:05:01.753797] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:44.293 [2024-12-05 19:05:01.753841] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:44.293 [2024-12-05 19:05:01.753859] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:44.293 pt3 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.293 [2024-12-05 19:05:01.765400] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:44.293 [2024-12-05 19:05:01.765451] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:44.293 [2024-12-05 19:05:01.765462] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:14:44.293 [2024-12-05 19:05:01.765472] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:44.293 [2024-12-05 19:05:01.765755] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:44.293 [2024-12-05 19:05:01.765773] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:44.293 [2024-12-05 19:05:01.765817] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:14:44.293 [2024-12-05 19:05:01.765834] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:44.293 [2024-12-05 19:05:01.765921] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:44.293 [2024-12-05 19:05:01.765932] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:44.293 [2024-12-05 19:05:01.766160] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:14:44.293 [2024-12-05 19:05:01.766611] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:44.293 [2024-12-05 19:05:01.766627] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:14:44.293 [2024-12-05 19:05:01.766729] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:44.293 pt4 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.293 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:44.293 "name": "raid_bdev1", 00:14:44.293 "uuid": "0af56e1f-2d1a-4451-9269-e054d1ec2256", 00:14:44.293 "strip_size_kb": 64, 00:14:44.293 "state": "online", 00:14:44.293 "raid_level": "raid5f", 00:14:44.293 "superblock": true, 00:14:44.293 "num_base_bdevs": 4, 00:14:44.293 "num_base_bdevs_discovered": 4, 00:14:44.293 "num_base_bdevs_operational": 4, 00:14:44.293 "base_bdevs_list": [ 00:14:44.293 { 00:14:44.293 "name": "pt1", 00:14:44.293 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:44.293 "is_configured": true, 00:14:44.293 "data_offset": 2048, 00:14:44.293 "data_size": 63488 00:14:44.293 }, 00:14:44.293 { 00:14:44.293 "name": "pt2", 00:14:44.293 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:44.293 "is_configured": true, 00:14:44.293 "data_offset": 2048, 00:14:44.293 "data_size": 63488 00:14:44.293 }, 00:14:44.293 { 00:14:44.293 "name": "pt3", 00:14:44.293 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:44.293 "is_configured": true, 00:14:44.293 "data_offset": 2048, 00:14:44.293 "data_size": 63488 00:14:44.293 }, 00:14:44.293 { 00:14:44.293 "name": "pt4", 00:14:44.293 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:44.293 "is_configured": true, 00:14:44.293 "data_offset": 2048, 00:14:44.293 "data_size": 63488 00:14:44.294 } 00:14:44.294 ] 00:14:44.294 }' 00:14:44.294 19:05:01 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:44.294 19:05:01 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.863 [2024-12-05 19:05:02.172869] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:44.863 "name": "raid_bdev1", 00:14:44.863 "aliases": [ 00:14:44.863 "0af56e1f-2d1a-4451-9269-e054d1ec2256" 00:14:44.863 ], 00:14:44.863 "product_name": "Raid Volume", 00:14:44.863 "block_size": 512, 00:14:44.863 "num_blocks": 190464, 00:14:44.863 "uuid": "0af56e1f-2d1a-4451-9269-e054d1ec2256", 00:14:44.863 "assigned_rate_limits": { 00:14:44.863 "rw_ios_per_sec": 0, 00:14:44.863 "rw_mbytes_per_sec": 0, 00:14:44.863 "r_mbytes_per_sec": 0, 00:14:44.863 "w_mbytes_per_sec": 0 00:14:44.863 }, 00:14:44.863 "claimed": false, 00:14:44.863 "zoned": false, 00:14:44.863 "supported_io_types": { 00:14:44.863 "read": true, 00:14:44.863 "write": true, 00:14:44.863 "unmap": false, 00:14:44.863 "flush": false, 00:14:44.863 "reset": true, 00:14:44.863 "nvme_admin": false, 00:14:44.863 "nvme_io": false, 00:14:44.863 "nvme_io_md": false, 00:14:44.863 "write_zeroes": true, 00:14:44.863 "zcopy": false, 00:14:44.863 "get_zone_info": false, 00:14:44.863 "zone_management": false, 00:14:44.863 "zone_append": false, 00:14:44.863 "compare": false, 00:14:44.863 "compare_and_write": false, 00:14:44.863 "abort": false, 00:14:44.863 "seek_hole": false, 00:14:44.863 "seek_data": false, 00:14:44.863 "copy": false, 00:14:44.863 "nvme_iov_md": false 00:14:44.863 }, 00:14:44.863 "driver_specific": { 00:14:44.863 "raid": { 00:14:44.863 "uuid": "0af56e1f-2d1a-4451-9269-e054d1ec2256", 00:14:44.863 "strip_size_kb": 64, 00:14:44.863 "state": "online", 00:14:44.863 "raid_level": "raid5f", 00:14:44.863 "superblock": true, 00:14:44.863 "num_base_bdevs": 4, 00:14:44.863 "num_base_bdevs_discovered": 4, 00:14:44.863 "num_base_bdevs_operational": 4, 00:14:44.863 "base_bdevs_list": [ 00:14:44.863 { 00:14:44.863 "name": "pt1", 00:14:44.863 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:44.863 "is_configured": true, 00:14:44.863 "data_offset": 2048, 00:14:44.863 "data_size": 63488 00:14:44.863 }, 00:14:44.863 { 00:14:44.863 "name": "pt2", 00:14:44.863 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:44.863 "is_configured": true, 00:14:44.863 "data_offset": 2048, 00:14:44.863 "data_size": 63488 00:14:44.863 }, 00:14:44.863 { 00:14:44.863 "name": "pt3", 00:14:44.863 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:44.863 "is_configured": true, 00:14:44.863 "data_offset": 2048, 00:14:44.863 "data_size": 63488 00:14:44.863 }, 00:14:44.863 { 00:14:44.863 "name": "pt4", 00:14:44.863 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:44.863 "is_configured": true, 00:14:44.863 "data_offset": 2048, 00:14:44.863 "data_size": 63488 00:14:44.863 } 00:14:44.863 ] 00:14:44.863 } 00:14:44.863 } 00:14:44.863 }' 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:44.863 pt2 00:14:44.863 pt3 00:14:44.863 pt4' 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:44.863 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:44.864 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:44.864 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:44.864 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:44.864 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:44.864 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.864 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:44.864 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.122 [2024-12-05 19:05:02.484303] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 0af56e1f-2d1a-4451-9269-e054d1ec2256 '!=' 0af56e1f-2d1a-4451-9269-e054d1ec2256 ']' 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.122 [2024-12-05 19:05:02.512133] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.122 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:45.122 "name": "raid_bdev1", 00:14:45.122 "uuid": "0af56e1f-2d1a-4451-9269-e054d1ec2256", 00:14:45.122 "strip_size_kb": 64, 00:14:45.122 "state": "online", 00:14:45.122 "raid_level": "raid5f", 00:14:45.122 "superblock": true, 00:14:45.122 "num_base_bdevs": 4, 00:14:45.122 "num_base_bdevs_discovered": 3, 00:14:45.122 "num_base_bdevs_operational": 3, 00:14:45.122 "base_bdevs_list": [ 00:14:45.123 { 00:14:45.123 "name": null, 00:14:45.123 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:45.123 "is_configured": false, 00:14:45.123 "data_offset": 0, 00:14:45.123 "data_size": 63488 00:14:45.123 }, 00:14:45.123 { 00:14:45.123 "name": "pt2", 00:14:45.123 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:45.123 "is_configured": true, 00:14:45.123 "data_offset": 2048, 00:14:45.123 "data_size": 63488 00:14:45.123 }, 00:14:45.123 { 00:14:45.123 "name": "pt3", 00:14:45.123 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:45.123 "is_configured": true, 00:14:45.123 "data_offset": 2048, 00:14:45.123 "data_size": 63488 00:14:45.123 }, 00:14:45.123 { 00:14:45.123 "name": "pt4", 00:14:45.123 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:45.123 "is_configured": true, 00:14:45.123 "data_offset": 2048, 00:14:45.123 "data_size": 63488 00:14:45.123 } 00:14:45.123 ] 00:14:45.123 }' 00:14:45.123 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:45.123 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.382 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:45.382 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.382 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.382 [2024-12-05 19:05:02.895609] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:45.382 [2024-12-05 19:05:02.895692] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:45.382 [2024-12-05 19:05:02.895774] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:45.382 [2024-12-05 19:05:02.895860] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:45.382 [2024-12-05 19:05:02.895927] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:14:45.382 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.382 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.382 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:14:45.382 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.382 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.382 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.642 [2024-12-05 19:05:02.991438] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:45.642 [2024-12-05 19:05:02.991528] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:45.642 [2024-12-05 19:05:02.991572] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:14:45.642 [2024-12-05 19:05:02.991599] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:45.642 [2024-12-05 19:05:02.993624] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:45.642 [2024-12-05 19:05:02.993718] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:45.642 [2024-12-05 19:05:02.993793] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:45.642 [2024-12-05 19:05:02.993848] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:45.642 pt2 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:45.642 19:05:02 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:45.642 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.642 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:45.642 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.642 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.643 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.643 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:45.643 "name": "raid_bdev1", 00:14:45.643 "uuid": "0af56e1f-2d1a-4451-9269-e054d1ec2256", 00:14:45.643 "strip_size_kb": 64, 00:14:45.643 "state": "configuring", 00:14:45.643 "raid_level": "raid5f", 00:14:45.643 "superblock": true, 00:14:45.643 "num_base_bdevs": 4, 00:14:45.643 "num_base_bdevs_discovered": 1, 00:14:45.643 "num_base_bdevs_operational": 3, 00:14:45.643 "base_bdevs_list": [ 00:14:45.643 { 00:14:45.643 "name": null, 00:14:45.643 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:45.643 "is_configured": false, 00:14:45.643 "data_offset": 2048, 00:14:45.643 "data_size": 63488 00:14:45.643 }, 00:14:45.643 { 00:14:45.643 "name": "pt2", 00:14:45.643 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:45.643 "is_configured": true, 00:14:45.643 "data_offset": 2048, 00:14:45.643 "data_size": 63488 00:14:45.643 }, 00:14:45.643 { 00:14:45.643 "name": null, 00:14:45.643 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:45.643 "is_configured": false, 00:14:45.643 "data_offset": 2048, 00:14:45.643 "data_size": 63488 00:14:45.643 }, 00:14:45.643 { 00:14:45.643 "name": null, 00:14:45.643 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:45.643 "is_configured": false, 00:14:45.643 "data_offset": 2048, 00:14:45.643 "data_size": 63488 00:14:45.643 } 00:14:45.643 ] 00:14:45.643 }' 00:14:45.643 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:45.643 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.903 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:14:45.903 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:45.903 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:45.903 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.903 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.903 [2024-12-05 19:05:03.418759] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:45.903 [2024-12-05 19:05:03.418856] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:45.903 [2024-12-05 19:05:03.418901] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:14:45.903 [2024-12-05 19:05:03.418929] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:45.903 [2024-12-05 19:05:03.419229] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:45.903 [2024-12-05 19:05:03.419292] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:45.903 [2024-12-05 19:05:03.419369] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:45.903 [2024-12-05 19:05:03.419420] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:45.903 pt3 00:14:45.903 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:45.903 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:45.903 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:45.903 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:45.903 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:45.903 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:45.903 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:45.903 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:45.903 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:45.904 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:45.904 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:45.904 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.904 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:45.904 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.904 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:45.904 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.163 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:46.163 "name": "raid_bdev1", 00:14:46.163 "uuid": "0af56e1f-2d1a-4451-9269-e054d1ec2256", 00:14:46.163 "strip_size_kb": 64, 00:14:46.163 "state": "configuring", 00:14:46.163 "raid_level": "raid5f", 00:14:46.163 "superblock": true, 00:14:46.163 "num_base_bdevs": 4, 00:14:46.163 "num_base_bdevs_discovered": 2, 00:14:46.163 "num_base_bdevs_operational": 3, 00:14:46.163 "base_bdevs_list": [ 00:14:46.163 { 00:14:46.163 "name": null, 00:14:46.163 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.163 "is_configured": false, 00:14:46.163 "data_offset": 2048, 00:14:46.163 "data_size": 63488 00:14:46.163 }, 00:14:46.163 { 00:14:46.163 "name": "pt2", 00:14:46.163 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:46.163 "is_configured": true, 00:14:46.163 "data_offset": 2048, 00:14:46.163 "data_size": 63488 00:14:46.163 }, 00:14:46.163 { 00:14:46.163 "name": "pt3", 00:14:46.163 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:46.163 "is_configured": true, 00:14:46.163 "data_offset": 2048, 00:14:46.163 "data_size": 63488 00:14:46.163 }, 00:14:46.163 { 00:14:46.163 "name": null, 00:14:46.163 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:46.163 "is_configured": false, 00:14:46.163 "data_offset": 2048, 00:14:46.163 "data_size": 63488 00:14:46.163 } 00:14:46.163 ] 00:14:46.163 }' 00:14:46.163 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:46.163 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.422 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:14:46.422 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:46.422 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:14:46.422 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:46.422 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.423 [2024-12-05 19:05:03.861949] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:46.423 [2024-12-05 19:05:03.862043] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:46.423 [2024-12-05 19:05:03.862088] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:14:46.423 [2024-12-05 19:05:03.862115] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:46.423 [2024-12-05 19:05:03.862427] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:46.423 [2024-12-05 19:05:03.862485] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:46.423 [2024-12-05 19:05:03.862559] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:14:46.423 [2024-12-05 19:05:03.862605] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:46.423 [2024-12-05 19:05:03.862715] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:46.423 [2024-12-05 19:05:03.862757] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:46.423 [2024-12-05 19:05:03.862980] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:14:46.423 [2024-12-05 19:05:03.863504] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:46.423 [2024-12-05 19:05:03.863554] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:14:46.423 [2024-12-05 19:05:03.863783] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:46.423 pt4 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:46.423 "name": "raid_bdev1", 00:14:46.423 "uuid": "0af56e1f-2d1a-4451-9269-e054d1ec2256", 00:14:46.423 "strip_size_kb": 64, 00:14:46.423 "state": "online", 00:14:46.423 "raid_level": "raid5f", 00:14:46.423 "superblock": true, 00:14:46.423 "num_base_bdevs": 4, 00:14:46.423 "num_base_bdevs_discovered": 3, 00:14:46.423 "num_base_bdevs_operational": 3, 00:14:46.423 "base_bdevs_list": [ 00:14:46.423 { 00:14:46.423 "name": null, 00:14:46.423 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.423 "is_configured": false, 00:14:46.423 "data_offset": 2048, 00:14:46.423 "data_size": 63488 00:14:46.423 }, 00:14:46.423 { 00:14:46.423 "name": "pt2", 00:14:46.423 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:46.423 "is_configured": true, 00:14:46.423 "data_offset": 2048, 00:14:46.423 "data_size": 63488 00:14:46.423 }, 00:14:46.423 { 00:14:46.423 "name": "pt3", 00:14:46.423 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:46.423 "is_configured": true, 00:14:46.423 "data_offset": 2048, 00:14:46.423 "data_size": 63488 00:14:46.423 }, 00:14:46.423 { 00:14:46.423 "name": "pt4", 00:14:46.423 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:46.423 "is_configured": true, 00:14:46.423 "data_offset": 2048, 00:14:46.423 "data_size": 63488 00:14:46.423 } 00:14:46.423 ] 00:14:46.423 }' 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:46.423 19:05:03 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.993 [2024-12-05 19:05:04.277356] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:46.993 [2024-12-05 19:05:04.277432] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:46.993 [2024-12-05 19:05:04.277511] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:46.993 [2024-12-05 19:05:04.277573] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:46.993 [2024-12-05 19:05:04.277582] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.993 [2024-12-05 19:05:04.353241] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:46.993 [2024-12-05 19:05:04.353339] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:46.993 [2024-12-05 19:05:04.353390] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:14:46.993 [2024-12-05 19:05:04.353416] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:46.993 [2024-12-05 19:05:04.355460] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:46.993 [2024-12-05 19:05:04.355543] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:46.993 [2024-12-05 19:05:04.355622] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:46.993 [2024-12-05 19:05:04.355669] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:46.993 [2024-12-05 19:05:04.355791] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:14:46.993 [2024-12-05 19:05:04.355851] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:46.993 [2024-12-05 19:05:04.355941] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:14:46.993 [2024-12-05 19:05:04.356056] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:46.993 [2024-12-05 19:05:04.356202] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:46.993 pt1 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:46.993 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:46.993 "name": "raid_bdev1", 00:14:46.993 "uuid": "0af56e1f-2d1a-4451-9269-e054d1ec2256", 00:14:46.993 "strip_size_kb": 64, 00:14:46.993 "state": "configuring", 00:14:46.993 "raid_level": "raid5f", 00:14:46.993 "superblock": true, 00:14:46.993 "num_base_bdevs": 4, 00:14:46.993 "num_base_bdevs_discovered": 2, 00:14:46.993 "num_base_bdevs_operational": 3, 00:14:46.993 "base_bdevs_list": [ 00:14:46.993 { 00:14:46.993 "name": null, 00:14:46.994 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.994 "is_configured": false, 00:14:46.994 "data_offset": 2048, 00:14:46.994 "data_size": 63488 00:14:46.994 }, 00:14:46.994 { 00:14:46.994 "name": "pt2", 00:14:46.994 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:46.994 "is_configured": true, 00:14:46.994 "data_offset": 2048, 00:14:46.994 "data_size": 63488 00:14:46.994 }, 00:14:46.994 { 00:14:46.994 "name": "pt3", 00:14:46.994 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:46.994 "is_configured": true, 00:14:46.994 "data_offset": 2048, 00:14:46.994 "data_size": 63488 00:14:46.994 }, 00:14:46.994 { 00:14:46.994 "name": null, 00:14:46.994 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:46.994 "is_configured": false, 00:14:46.994 "data_offset": 2048, 00:14:46.994 "data_size": 63488 00:14:46.994 } 00:14:46.994 ] 00:14:46.994 }' 00:14:46.994 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:46.994 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.253 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:14:47.253 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:14:47.253 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.253 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.513 [2024-12-05 19:05:04.836563] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:47.513 [2024-12-05 19:05:04.836690] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:47.513 [2024-12-05 19:05:04.836730] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:14:47.513 [2024-12-05 19:05:04.836760] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:47.513 [2024-12-05 19:05:04.837070] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:47.513 [2024-12-05 19:05:04.837129] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:47.513 [2024-12-05 19:05:04.837205] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:14:47.513 [2024-12-05 19:05:04.837258] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:47.513 [2024-12-05 19:05:04.837357] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:14:47.513 [2024-12-05 19:05:04.837398] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:47.513 [2024-12-05 19:05:04.837641] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:14:47.513 [2024-12-05 19:05:04.838181] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:14:47.513 [2024-12-05 19:05:04.838238] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:14:47.513 [2024-12-05 19:05:04.838437] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:47.513 pt4 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:47.513 "name": "raid_bdev1", 00:14:47.513 "uuid": "0af56e1f-2d1a-4451-9269-e054d1ec2256", 00:14:47.513 "strip_size_kb": 64, 00:14:47.513 "state": "online", 00:14:47.513 "raid_level": "raid5f", 00:14:47.513 "superblock": true, 00:14:47.513 "num_base_bdevs": 4, 00:14:47.513 "num_base_bdevs_discovered": 3, 00:14:47.513 "num_base_bdevs_operational": 3, 00:14:47.513 "base_bdevs_list": [ 00:14:47.513 { 00:14:47.513 "name": null, 00:14:47.513 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.513 "is_configured": false, 00:14:47.513 "data_offset": 2048, 00:14:47.513 "data_size": 63488 00:14:47.513 }, 00:14:47.513 { 00:14:47.513 "name": "pt2", 00:14:47.513 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:47.513 "is_configured": true, 00:14:47.513 "data_offset": 2048, 00:14:47.513 "data_size": 63488 00:14:47.513 }, 00:14:47.513 { 00:14:47.513 "name": "pt3", 00:14:47.513 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:47.513 "is_configured": true, 00:14:47.513 "data_offset": 2048, 00:14:47.513 "data_size": 63488 00:14:47.513 }, 00:14:47.513 { 00:14:47.513 "name": "pt4", 00:14:47.513 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:47.513 "is_configured": true, 00:14:47.513 "data_offset": 2048, 00:14:47.513 "data_size": 63488 00:14:47.513 } 00:14:47.513 ] 00:14:47.513 }' 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:47.513 19:05:04 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.772 19:05:05 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:14:47.772 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:47.772 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.772 19:05:05 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:14:47.773 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.032 [2024-12-05 19:05:05.343893] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 0af56e1f-2d1a-4451-9269-e054d1ec2256 '!=' 0af56e1f-2d1a-4451-9269-e054d1ec2256 ']' 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 92042 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # '[' -z 92042 ']' 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@958 -- # kill -0 92042 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # uname 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 92042 00:14:48.032 killing process with pid 92042 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 92042' 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@973 -- # kill 92042 00:14:48.032 [2024-12-05 19:05:05.418815] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:48.032 [2024-12-05 19:05:05.418877] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:48.032 [2024-12-05 19:05:05.418935] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:48.032 [2024-12-05 19:05:05.418944] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:14:48.032 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@978 -- # wait 92042 00:14:48.032 [2024-12-05 19:05:05.461165] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:48.290 19:05:05 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:14:48.290 ************************************ 00:14:48.290 END TEST raid5f_superblock_test 00:14:48.290 ************************************ 00:14:48.290 00:14:48.290 real 0m6.912s 00:14:48.290 user 0m11.615s 00:14:48.290 sys 0m1.530s 00:14:48.290 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:48.290 19:05:05 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.290 19:05:05 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:14:48.290 19:05:05 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 4 false false true 00:14:48.290 19:05:05 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:14:48.290 19:05:05 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:48.290 19:05:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:48.290 ************************************ 00:14:48.290 START TEST raid5f_rebuild_test 00:14:48.290 ************************************ 00:14:48.290 19:05:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 4 false false true 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=92470 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 92470 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # '[' -z 92470 ']' 00:14:48.291 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:48.291 19:05:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:48.550 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:48.550 Zero copy mechanism will not be used. 00:14:48.550 [2024-12-05 19:05:05.863906] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:14:48.550 [2024-12-05 19:05:05.864075] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92470 ] 00:14:48.550 [2024-12-05 19:05:06.020756] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:48.550 [2024-12-05 19:05:06.044397] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:48.550 [2024-12-05 19:05:06.087729] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:48.550 [2024-12-05 19:05:06.087765] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:49.153 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:49.153 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # return 0 00:14:49.153 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:49.153 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:49.153 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.153 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.153 BaseBdev1_malloc 00:14:49.153 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.153 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:49.153 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.153 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.432 [2024-12-05 19:05:06.699567] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:49.432 [2024-12-05 19:05:06.699724] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:49.432 [2024-12-05 19:05:06.699783] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:49.432 [2024-12-05 19:05:06.699815] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:49.432 [2024-12-05 19:05:06.701885] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:49.432 [2024-12-05 19:05:06.701956] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:49.432 BaseBdev1 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.432 BaseBdev2_malloc 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.432 [2024-12-05 19:05:06.728104] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:49.432 [2024-12-05 19:05:06.728216] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:49.432 [2024-12-05 19:05:06.728259] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:49.432 [2024-12-05 19:05:06.728267] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:49.432 [2024-12-05 19:05:06.730275] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:49.432 [2024-12-05 19:05:06.730317] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:49.432 BaseBdev2 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.432 BaseBdev3_malloc 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.432 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.432 [2024-12-05 19:05:06.756619] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:49.432 [2024-12-05 19:05:06.756736] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:49.432 [2024-12-05 19:05:06.756792] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:49.432 [2024-12-05 19:05:06.756820] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:49.432 [2024-12-05 19:05:06.758842] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:49.432 [2024-12-05 19:05:06.758906] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:49.433 BaseBdev3 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.433 BaseBdev4_malloc 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.433 [2024-12-05 19:05:06.799008] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:14:49.433 [2024-12-05 19:05:06.799138] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:49.433 [2024-12-05 19:05:06.799173] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:49.433 [2024-12-05 19:05:06.799185] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:49.433 [2024-12-05 19:05:06.801821] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:49.433 BaseBdev4 00:14:49.433 [2024-12-05 19:05:06.801918] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.433 spare_malloc 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.433 spare_delay 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.433 [2024-12-05 19:05:06.839476] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:49.433 [2024-12-05 19:05:06.839577] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:49.433 [2024-12-05 19:05:06.839628] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:14:49.433 [2024-12-05 19:05:06.839656] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:49.433 [2024-12-05 19:05:06.841630] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:49.433 [2024-12-05 19:05:06.841710] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:49.433 spare 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.433 [2024-12-05 19:05:06.851533] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:49.433 [2024-12-05 19:05:06.853334] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:49.433 [2024-12-05 19:05:06.853393] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:49.433 [2024-12-05 19:05:06.853436] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:49.433 [2024-12-05 19:05:06.853520] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:49.433 [2024-12-05 19:05:06.853529] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:49.433 [2024-12-05 19:05:06.853770] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:14:49.433 [2024-12-05 19:05:06.854184] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:49.433 [2024-12-05 19:05:06.854197] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:49.433 [2024-12-05 19:05:06.854296] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:49.433 "name": "raid_bdev1", 00:14:49.433 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:14:49.433 "strip_size_kb": 64, 00:14:49.433 "state": "online", 00:14:49.433 "raid_level": "raid5f", 00:14:49.433 "superblock": false, 00:14:49.433 "num_base_bdevs": 4, 00:14:49.433 "num_base_bdevs_discovered": 4, 00:14:49.433 "num_base_bdevs_operational": 4, 00:14:49.433 "base_bdevs_list": [ 00:14:49.433 { 00:14:49.433 "name": "BaseBdev1", 00:14:49.433 "uuid": "5127565e-34ab-51b3-910f-e1353c7c190c", 00:14:49.433 "is_configured": true, 00:14:49.433 "data_offset": 0, 00:14:49.433 "data_size": 65536 00:14:49.433 }, 00:14:49.433 { 00:14:49.433 "name": "BaseBdev2", 00:14:49.433 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:14:49.433 "is_configured": true, 00:14:49.433 "data_offset": 0, 00:14:49.433 "data_size": 65536 00:14:49.433 }, 00:14:49.433 { 00:14:49.433 "name": "BaseBdev3", 00:14:49.433 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:14:49.433 "is_configured": true, 00:14:49.433 "data_offset": 0, 00:14:49.433 "data_size": 65536 00:14:49.433 }, 00:14:49.433 { 00:14:49.433 "name": "BaseBdev4", 00:14:49.433 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:14:49.433 "is_configured": true, 00:14:49.433 "data_offset": 0, 00:14:49.433 "data_size": 65536 00:14:49.433 } 00:14:49.433 ] 00:14:49.433 }' 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:49.433 19:05:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.003 [2024-12-05 19:05:07.327368] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=196608 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:50.003 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:50.264 [2024-12-05 19:05:07.574791] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:14:50.264 /dev/nbd0 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:50.264 1+0 records in 00:14:50.264 1+0 records out 00:14:50.264 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000311425 s, 13.2 MB/s 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 192 00:14:50.264 19:05:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=512 oflag=direct 00:14:50.525 512+0 records in 00:14:50.525 512+0 records out 00:14:50.525 100663296 bytes (101 MB, 96 MiB) copied, 0.377759 s, 266 MB/s 00:14:50.525 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:50.525 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:50.525 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:50.525 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:50.525 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:14:50.525 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:50.525 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:50.786 [2024-12-05 19:05:08.241728] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.786 [2024-12-05 19:05:08.253792] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:50.786 "name": "raid_bdev1", 00:14:50.786 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:14:50.786 "strip_size_kb": 64, 00:14:50.786 "state": "online", 00:14:50.786 "raid_level": "raid5f", 00:14:50.786 "superblock": false, 00:14:50.786 "num_base_bdevs": 4, 00:14:50.786 "num_base_bdevs_discovered": 3, 00:14:50.786 "num_base_bdevs_operational": 3, 00:14:50.786 "base_bdevs_list": [ 00:14:50.786 { 00:14:50.786 "name": null, 00:14:50.786 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.786 "is_configured": false, 00:14:50.786 "data_offset": 0, 00:14:50.786 "data_size": 65536 00:14:50.786 }, 00:14:50.786 { 00:14:50.786 "name": "BaseBdev2", 00:14:50.786 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:14:50.786 "is_configured": true, 00:14:50.786 "data_offset": 0, 00:14:50.786 "data_size": 65536 00:14:50.786 }, 00:14:50.786 { 00:14:50.786 "name": "BaseBdev3", 00:14:50.786 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:14:50.786 "is_configured": true, 00:14:50.786 "data_offset": 0, 00:14:50.786 "data_size": 65536 00:14:50.786 }, 00:14:50.786 { 00:14:50.786 "name": "BaseBdev4", 00:14:50.786 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:14:50.786 "is_configured": true, 00:14:50.786 "data_offset": 0, 00:14:50.786 "data_size": 65536 00:14:50.786 } 00:14:50.786 ] 00:14:50.786 }' 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:50.786 19:05:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.356 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:51.356 19:05:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:51.356 19:05:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.356 [2024-12-05 19:05:08.724955] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:51.356 [2024-12-05 19:05:08.729168] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027da0 00:14:51.356 19:05:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:51.356 19:05:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:51.356 [2024-12-05 19:05:08.731381] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:52.297 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:52.297 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:52.297 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:52.297 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:52.297 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:52.297 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.297 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:52.297 19:05:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.297 19:05:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.297 19:05:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.297 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:52.297 "name": "raid_bdev1", 00:14:52.297 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:14:52.297 "strip_size_kb": 64, 00:14:52.297 "state": "online", 00:14:52.297 "raid_level": "raid5f", 00:14:52.297 "superblock": false, 00:14:52.297 "num_base_bdevs": 4, 00:14:52.297 "num_base_bdevs_discovered": 4, 00:14:52.297 "num_base_bdevs_operational": 4, 00:14:52.297 "process": { 00:14:52.297 "type": "rebuild", 00:14:52.297 "target": "spare", 00:14:52.297 "progress": { 00:14:52.297 "blocks": 19200, 00:14:52.297 "percent": 9 00:14:52.297 } 00:14:52.297 }, 00:14:52.297 "base_bdevs_list": [ 00:14:52.297 { 00:14:52.297 "name": "spare", 00:14:52.297 "uuid": "0af1829b-866f-5890-a8a2-14229f30a473", 00:14:52.297 "is_configured": true, 00:14:52.297 "data_offset": 0, 00:14:52.297 "data_size": 65536 00:14:52.297 }, 00:14:52.297 { 00:14:52.297 "name": "BaseBdev2", 00:14:52.297 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:14:52.297 "is_configured": true, 00:14:52.297 "data_offset": 0, 00:14:52.297 "data_size": 65536 00:14:52.297 }, 00:14:52.297 { 00:14:52.297 "name": "BaseBdev3", 00:14:52.297 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:14:52.297 "is_configured": true, 00:14:52.297 "data_offset": 0, 00:14:52.297 "data_size": 65536 00:14:52.297 }, 00:14:52.297 { 00:14:52.297 "name": "BaseBdev4", 00:14:52.297 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:14:52.297 "is_configured": true, 00:14:52.297 "data_offset": 0, 00:14:52.297 "data_size": 65536 00:14:52.297 } 00:14:52.297 ] 00:14:52.297 }' 00:14:52.297 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:52.297 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:52.297 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.556 [2024-12-05 19:05:09.892062] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:52.556 [2024-12-05 19:05:09.936505] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:52.556 [2024-12-05 19:05:09.936610] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:52.556 [2024-12-05 19:05:09.936679] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:52.556 [2024-12-05 19:05:09.936703] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.556 19:05:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:52.557 19:05:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.557 19:05:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.557 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:52.557 "name": "raid_bdev1", 00:14:52.557 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:14:52.557 "strip_size_kb": 64, 00:14:52.557 "state": "online", 00:14:52.557 "raid_level": "raid5f", 00:14:52.557 "superblock": false, 00:14:52.557 "num_base_bdevs": 4, 00:14:52.557 "num_base_bdevs_discovered": 3, 00:14:52.557 "num_base_bdevs_operational": 3, 00:14:52.557 "base_bdevs_list": [ 00:14:52.557 { 00:14:52.557 "name": null, 00:14:52.557 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:52.557 "is_configured": false, 00:14:52.557 "data_offset": 0, 00:14:52.557 "data_size": 65536 00:14:52.557 }, 00:14:52.557 { 00:14:52.557 "name": "BaseBdev2", 00:14:52.557 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:14:52.557 "is_configured": true, 00:14:52.557 "data_offset": 0, 00:14:52.557 "data_size": 65536 00:14:52.557 }, 00:14:52.557 { 00:14:52.557 "name": "BaseBdev3", 00:14:52.557 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:14:52.557 "is_configured": true, 00:14:52.557 "data_offset": 0, 00:14:52.557 "data_size": 65536 00:14:52.557 }, 00:14:52.557 { 00:14:52.557 "name": "BaseBdev4", 00:14:52.557 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:14:52.557 "is_configured": true, 00:14:52.557 "data_offset": 0, 00:14:52.557 "data_size": 65536 00:14:52.557 } 00:14:52.557 ] 00:14:52.557 }' 00:14:52.557 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:52.557 19:05:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.818 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:52.818 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:52.818 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:52.818 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:52.818 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:52.818 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.818 19:05:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:52.818 19:05:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.818 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:52.818 19:05:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:52.818 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:52.818 "name": "raid_bdev1", 00:14:52.818 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:14:52.818 "strip_size_kb": 64, 00:14:52.818 "state": "online", 00:14:52.818 "raid_level": "raid5f", 00:14:52.818 "superblock": false, 00:14:52.818 "num_base_bdevs": 4, 00:14:52.818 "num_base_bdevs_discovered": 3, 00:14:52.818 "num_base_bdevs_operational": 3, 00:14:52.818 "base_bdevs_list": [ 00:14:52.818 { 00:14:52.818 "name": null, 00:14:52.818 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:52.818 "is_configured": false, 00:14:52.818 "data_offset": 0, 00:14:52.818 "data_size": 65536 00:14:52.818 }, 00:14:52.818 { 00:14:52.818 "name": "BaseBdev2", 00:14:52.818 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:14:52.818 "is_configured": true, 00:14:52.818 "data_offset": 0, 00:14:52.818 "data_size": 65536 00:14:52.818 }, 00:14:52.818 { 00:14:52.818 "name": "BaseBdev3", 00:14:52.818 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:14:52.818 "is_configured": true, 00:14:52.818 "data_offset": 0, 00:14:52.818 "data_size": 65536 00:14:52.818 }, 00:14:52.818 { 00:14:52.818 "name": "BaseBdev4", 00:14:52.818 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:14:52.818 "is_configured": true, 00:14:52.818 "data_offset": 0, 00:14:52.818 "data_size": 65536 00:14:52.818 } 00:14:52.818 ] 00:14:52.818 }' 00:14:52.818 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:53.078 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:53.078 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:53.078 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:53.078 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:53.078 19:05:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:53.078 19:05:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.078 [2024-12-05 19:05:10.429138] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:53.078 [2024-12-05 19:05:10.432242] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027e70 00:14:53.078 [2024-12-05 19:05:10.434410] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:53.078 19:05:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:53.078 19:05:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:54.019 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:54.019 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:54.019 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:54.019 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:54.019 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:54.019 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.019 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:54.019 19:05:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.019 19:05:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.019 19:05:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.019 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:54.019 "name": "raid_bdev1", 00:14:54.019 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:14:54.019 "strip_size_kb": 64, 00:14:54.019 "state": "online", 00:14:54.019 "raid_level": "raid5f", 00:14:54.019 "superblock": false, 00:14:54.019 "num_base_bdevs": 4, 00:14:54.019 "num_base_bdevs_discovered": 4, 00:14:54.019 "num_base_bdevs_operational": 4, 00:14:54.019 "process": { 00:14:54.019 "type": "rebuild", 00:14:54.019 "target": "spare", 00:14:54.019 "progress": { 00:14:54.019 "blocks": 19200, 00:14:54.019 "percent": 9 00:14:54.019 } 00:14:54.019 }, 00:14:54.019 "base_bdevs_list": [ 00:14:54.019 { 00:14:54.019 "name": "spare", 00:14:54.019 "uuid": "0af1829b-866f-5890-a8a2-14229f30a473", 00:14:54.019 "is_configured": true, 00:14:54.019 "data_offset": 0, 00:14:54.019 "data_size": 65536 00:14:54.019 }, 00:14:54.019 { 00:14:54.019 "name": "BaseBdev2", 00:14:54.019 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:14:54.019 "is_configured": true, 00:14:54.019 "data_offset": 0, 00:14:54.019 "data_size": 65536 00:14:54.019 }, 00:14:54.019 { 00:14:54.019 "name": "BaseBdev3", 00:14:54.019 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:14:54.019 "is_configured": true, 00:14:54.019 "data_offset": 0, 00:14:54.019 "data_size": 65536 00:14:54.019 }, 00:14:54.019 { 00:14:54.019 "name": "BaseBdev4", 00:14:54.019 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:14:54.019 "is_configured": true, 00:14:54.019 "data_offset": 0, 00:14:54.019 "data_size": 65536 00:14:54.019 } 00:14:54.019 ] 00:14:54.019 }' 00:14:54.019 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:54.019 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:54.019 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=501 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:54.280 "name": "raid_bdev1", 00:14:54.280 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:14:54.280 "strip_size_kb": 64, 00:14:54.280 "state": "online", 00:14:54.280 "raid_level": "raid5f", 00:14:54.280 "superblock": false, 00:14:54.280 "num_base_bdevs": 4, 00:14:54.280 "num_base_bdevs_discovered": 4, 00:14:54.280 "num_base_bdevs_operational": 4, 00:14:54.280 "process": { 00:14:54.280 "type": "rebuild", 00:14:54.280 "target": "spare", 00:14:54.280 "progress": { 00:14:54.280 "blocks": 21120, 00:14:54.280 "percent": 10 00:14:54.280 } 00:14:54.280 }, 00:14:54.280 "base_bdevs_list": [ 00:14:54.280 { 00:14:54.280 "name": "spare", 00:14:54.280 "uuid": "0af1829b-866f-5890-a8a2-14229f30a473", 00:14:54.280 "is_configured": true, 00:14:54.280 "data_offset": 0, 00:14:54.280 "data_size": 65536 00:14:54.280 }, 00:14:54.280 { 00:14:54.280 "name": "BaseBdev2", 00:14:54.280 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:14:54.280 "is_configured": true, 00:14:54.280 "data_offset": 0, 00:14:54.280 "data_size": 65536 00:14:54.280 }, 00:14:54.280 { 00:14:54.280 "name": "BaseBdev3", 00:14:54.280 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:14:54.280 "is_configured": true, 00:14:54.280 "data_offset": 0, 00:14:54.280 "data_size": 65536 00:14:54.280 }, 00:14:54.280 { 00:14:54.280 "name": "BaseBdev4", 00:14:54.280 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:14:54.280 "is_configured": true, 00:14:54.280 "data_offset": 0, 00:14:54.280 "data_size": 65536 00:14:54.280 } 00:14:54.280 ] 00:14:54.280 }' 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:54.280 19:05:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:55.221 19:05:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:55.221 19:05:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:55.221 19:05:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:55.221 19:05:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:55.221 19:05:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:55.221 19:05:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:55.221 19:05:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:55.221 19:05:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:55.221 19:05:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:55.221 19:05:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.221 19:05:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:55.481 19:05:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:55.481 "name": "raid_bdev1", 00:14:55.481 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:14:55.481 "strip_size_kb": 64, 00:14:55.481 "state": "online", 00:14:55.481 "raid_level": "raid5f", 00:14:55.481 "superblock": false, 00:14:55.481 "num_base_bdevs": 4, 00:14:55.481 "num_base_bdevs_discovered": 4, 00:14:55.481 "num_base_bdevs_operational": 4, 00:14:55.481 "process": { 00:14:55.481 "type": "rebuild", 00:14:55.481 "target": "spare", 00:14:55.481 "progress": { 00:14:55.481 "blocks": 44160, 00:14:55.481 "percent": 22 00:14:55.481 } 00:14:55.481 }, 00:14:55.481 "base_bdevs_list": [ 00:14:55.481 { 00:14:55.481 "name": "spare", 00:14:55.481 "uuid": "0af1829b-866f-5890-a8a2-14229f30a473", 00:14:55.481 "is_configured": true, 00:14:55.481 "data_offset": 0, 00:14:55.481 "data_size": 65536 00:14:55.481 }, 00:14:55.481 { 00:14:55.481 "name": "BaseBdev2", 00:14:55.481 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:14:55.481 "is_configured": true, 00:14:55.481 "data_offset": 0, 00:14:55.481 "data_size": 65536 00:14:55.481 }, 00:14:55.481 { 00:14:55.481 "name": "BaseBdev3", 00:14:55.481 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:14:55.481 "is_configured": true, 00:14:55.481 "data_offset": 0, 00:14:55.481 "data_size": 65536 00:14:55.481 }, 00:14:55.481 { 00:14:55.481 "name": "BaseBdev4", 00:14:55.481 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:14:55.481 "is_configured": true, 00:14:55.481 "data_offset": 0, 00:14:55.481 "data_size": 65536 00:14:55.481 } 00:14:55.481 ] 00:14:55.481 }' 00:14:55.481 19:05:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:55.481 19:05:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:55.481 19:05:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:55.481 19:05:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:55.481 19:05:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:56.420 19:05:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:56.420 19:05:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:56.420 19:05:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:56.420 19:05:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:56.420 19:05:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:56.420 19:05:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:56.420 19:05:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.420 19:05:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:56.420 19:05:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:56.420 19:05:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.420 19:05:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:56.421 19:05:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:56.421 "name": "raid_bdev1", 00:14:56.421 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:14:56.421 "strip_size_kb": 64, 00:14:56.421 "state": "online", 00:14:56.421 "raid_level": "raid5f", 00:14:56.421 "superblock": false, 00:14:56.421 "num_base_bdevs": 4, 00:14:56.421 "num_base_bdevs_discovered": 4, 00:14:56.421 "num_base_bdevs_operational": 4, 00:14:56.421 "process": { 00:14:56.421 "type": "rebuild", 00:14:56.421 "target": "spare", 00:14:56.421 "progress": { 00:14:56.421 "blocks": 65280, 00:14:56.421 "percent": 33 00:14:56.421 } 00:14:56.421 }, 00:14:56.421 "base_bdevs_list": [ 00:14:56.421 { 00:14:56.421 "name": "spare", 00:14:56.421 "uuid": "0af1829b-866f-5890-a8a2-14229f30a473", 00:14:56.421 "is_configured": true, 00:14:56.421 "data_offset": 0, 00:14:56.421 "data_size": 65536 00:14:56.421 }, 00:14:56.421 { 00:14:56.421 "name": "BaseBdev2", 00:14:56.421 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:14:56.421 "is_configured": true, 00:14:56.421 "data_offset": 0, 00:14:56.421 "data_size": 65536 00:14:56.421 }, 00:14:56.421 { 00:14:56.421 "name": "BaseBdev3", 00:14:56.421 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:14:56.421 "is_configured": true, 00:14:56.421 "data_offset": 0, 00:14:56.421 "data_size": 65536 00:14:56.421 }, 00:14:56.421 { 00:14:56.421 "name": "BaseBdev4", 00:14:56.421 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:14:56.421 "is_configured": true, 00:14:56.421 "data_offset": 0, 00:14:56.421 "data_size": 65536 00:14:56.421 } 00:14:56.421 ] 00:14:56.421 }' 00:14:56.421 19:05:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:56.681 19:05:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:56.681 19:05:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:56.681 19:05:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:56.681 19:05:14 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:57.622 19:05:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:57.623 19:05:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:57.623 19:05:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:57.623 19:05:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:57.623 19:05:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:57.623 19:05:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:57.623 19:05:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.623 19:05:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:57.623 19:05:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:57.623 19:05:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.623 19:05:15 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:57.623 19:05:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:57.623 "name": "raid_bdev1", 00:14:57.623 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:14:57.623 "strip_size_kb": 64, 00:14:57.623 "state": "online", 00:14:57.623 "raid_level": "raid5f", 00:14:57.623 "superblock": false, 00:14:57.623 "num_base_bdevs": 4, 00:14:57.623 "num_base_bdevs_discovered": 4, 00:14:57.623 "num_base_bdevs_operational": 4, 00:14:57.623 "process": { 00:14:57.623 "type": "rebuild", 00:14:57.623 "target": "spare", 00:14:57.623 "progress": { 00:14:57.623 "blocks": 86400, 00:14:57.623 "percent": 43 00:14:57.623 } 00:14:57.623 }, 00:14:57.623 "base_bdevs_list": [ 00:14:57.623 { 00:14:57.623 "name": "spare", 00:14:57.623 "uuid": "0af1829b-866f-5890-a8a2-14229f30a473", 00:14:57.623 "is_configured": true, 00:14:57.623 "data_offset": 0, 00:14:57.623 "data_size": 65536 00:14:57.623 }, 00:14:57.623 { 00:14:57.623 "name": "BaseBdev2", 00:14:57.623 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:14:57.623 "is_configured": true, 00:14:57.623 "data_offset": 0, 00:14:57.623 "data_size": 65536 00:14:57.623 }, 00:14:57.623 { 00:14:57.623 "name": "BaseBdev3", 00:14:57.623 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:14:57.623 "is_configured": true, 00:14:57.623 "data_offset": 0, 00:14:57.623 "data_size": 65536 00:14:57.623 }, 00:14:57.623 { 00:14:57.623 "name": "BaseBdev4", 00:14:57.623 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:14:57.623 "is_configured": true, 00:14:57.623 "data_offset": 0, 00:14:57.623 "data_size": 65536 00:14:57.623 } 00:14:57.623 ] 00:14:57.623 }' 00:14:57.623 19:05:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:57.623 19:05:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:57.623 19:05:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:57.882 19:05:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:57.882 19:05:15 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:58.820 "name": "raid_bdev1", 00:14:58.820 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:14:58.820 "strip_size_kb": 64, 00:14:58.820 "state": "online", 00:14:58.820 "raid_level": "raid5f", 00:14:58.820 "superblock": false, 00:14:58.820 "num_base_bdevs": 4, 00:14:58.820 "num_base_bdevs_discovered": 4, 00:14:58.820 "num_base_bdevs_operational": 4, 00:14:58.820 "process": { 00:14:58.820 "type": "rebuild", 00:14:58.820 "target": "spare", 00:14:58.820 "progress": { 00:14:58.820 "blocks": 109440, 00:14:58.820 "percent": 55 00:14:58.820 } 00:14:58.820 }, 00:14:58.820 "base_bdevs_list": [ 00:14:58.820 { 00:14:58.820 "name": "spare", 00:14:58.820 "uuid": "0af1829b-866f-5890-a8a2-14229f30a473", 00:14:58.820 "is_configured": true, 00:14:58.820 "data_offset": 0, 00:14:58.820 "data_size": 65536 00:14:58.820 }, 00:14:58.820 { 00:14:58.820 "name": "BaseBdev2", 00:14:58.820 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:14:58.820 "is_configured": true, 00:14:58.820 "data_offset": 0, 00:14:58.820 "data_size": 65536 00:14:58.820 }, 00:14:58.820 { 00:14:58.820 "name": "BaseBdev3", 00:14:58.820 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:14:58.820 "is_configured": true, 00:14:58.820 "data_offset": 0, 00:14:58.820 "data_size": 65536 00:14:58.820 }, 00:14:58.820 { 00:14:58.820 "name": "BaseBdev4", 00:14:58.820 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:14:58.820 "is_configured": true, 00:14:58.820 "data_offset": 0, 00:14:58.820 "data_size": 65536 00:14:58.820 } 00:14:58.820 ] 00:14:58.820 }' 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:58.820 19:05:16 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:00.202 "name": "raid_bdev1", 00:15:00.202 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:15:00.202 "strip_size_kb": 64, 00:15:00.202 "state": "online", 00:15:00.202 "raid_level": "raid5f", 00:15:00.202 "superblock": false, 00:15:00.202 "num_base_bdevs": 4, 00:15:00.202 "num_base_bdevs_discovered": 4, 00:15:00.202 "num_base_bdevs_operational": 4, 00:15:00.202 "process": { 00:15:00.202 "type": "rebuild", 00:15:00.202 "target": "spare", 00:15:00.202 "progress": { 00:15:00.202 "blocks": 130560, 00:15:00.202 "percent": 66 00:15:00.202 } 00:15:00.202 }, 00:15:00.202 "base_bdevs_list": [ 00:15:00.202 { 00:15:00.202 "name": "spare", 00:15:00.202 "uuid": "0af1829b-866f-5890-a8a2-14229f30a473", 00:15:00.202 "is_configured": true, 00:15:00.202 "data_offset": 0, 00:15:00.202 "data_size": 65536 00:15:00.202 }, 00:15:00.202 { 00:15:00.202 "name": "BaseBdev2", 00:15:00.202 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:15:00.202 "is_configured": true, 00:15:00.202 "data_offset": 0, 00:15:00.202 "data_size": 65536 00:15:00.202 }, 00:15:00.202 { 00:15:00.202 "name": "BaseBdev3", 00:15:00.202 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:15:00.202 "is_configured": true, 00:15:00.202 "data_offset": 0, 00:15:00.202 "data_size": 65536 00:15:00.202 }, 00:15:00.202 { 00:15:00.202 "name": "BaseBdev4", 00:15:00.202 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:15:00.202 "is_configured": true, 00:15:00.202 "data_offset": 0, 00:15:00.202 "data_size": 65536 00:15:00.202 } 00:15:00.202 ] 00:15:00.202 }' 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:00.202 19:05:17 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:01.141 19:05:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:01.141 19:05:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:01.141 19:05:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:01.141 19:05:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:01.141 19:05:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:01.141 19:05:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:01.141 19:05:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:01.141 19:05:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:01.142 19:05:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:01.142 19:05:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:01.142 19:05:18 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:01.142 19:05:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:01.142 "name": "raid_bdev1", 00:15:01.142 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:15:01.142 "strip_size_kb": 64, 00:15:01.142 "state": "online", 00:15:01.142 "raid_level": "raid5f", 00:15:01.142 "superblock": false, 00:15:01.142 "num_base_bdevs": 4, 00:15:01.142 "num_base_bdevs_discovered": 4, 00:15:01.142 "num_base_bdevs_operational": 4, 00:15:01.142 "process": { 00:15:01.142 "type": "rebuild", 00:15:01.142 "target": "spare", 00:15:01.142 "progress": { 00:15:01.142 "blocks": 153600, 00:15:01.142 "percent": 78 00:15:01.142 } 00:15:01.142 }, 00:15:01.142 "base_bdevs_list": [ 00:15:01.142 { 00:15:01.142 "name": "spare", 00:15:01.142 "uuid": "0af1829b-866f-5890-a8a2-14229f30a473", 00:15:01.142 "is_configured": true, 00:15:01.142 "data_offset": 0, 00:15:01.142 "data_size": 65536 00:15:01.142 }, 00:15:01.142 { 00:15:01.142 "name": "BaseBdev2", 00:15:01.142 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:15:01.142 "is_configured": true, 00:15:01.142 "data_offset": 0, 00:15:01.142 "data_size": 65536 00:15:01.142 }, 00:15:01.142 { 00:15:01.142 "name": "BaseBdev3", 00:15:01.142 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:15:01.142 "is_configured": true, 00:15:01.142 "data_offset": 0, 00:15:01.142 "data_size": 65536 00:15:01.142 }, 00:15:01.142 { 00:15:01.142 "name": "BaseBdev4", 00:15:01.142 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:15:01.142 "is_configured": true, 00:15:01.142 "data_offset": 0, 00:15:01.142 "data_size": 65536 00:15:01.142 } 00:15:01.142 ] 00:15:01.142 }' 00:15:01.142 19:05:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:01.142 19:05:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:01.142 19:05:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:01.142 19:05:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:01.142 19:05:18 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:02.079 19:05:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:02.079 19:05:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:02.079 19:05:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:02.079 19:05:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:02.079 19:05:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:02.079 19:05:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:02.338 19:05:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.338 19:05:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:02.338 19:05:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:02.338 19:05:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:02.338 19:05:19 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:02.338 19:05:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:02.338 "name": "raid_bdev1", 00:15:02.338 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:15:02.338 "strip_size_kb": 64, 00:15:02.338 "state": "online", 00:15:02.338 "raid_level": "raid5f", 00:15:02.338 "superblock": false, 00:15:02.338 "num_base_bdevs": 4, 00:15:02.338 "num_base_bdevs_discovered": 4, 00:15:02.338 "num_base_bdevs_operational": 4, 00:15:02.338 "process": { 00:15:02.338 "type": "rebuild", 00:15:02.338 "target": "spare", 00:15:02.338 "progress": { 00:15:02.338 "blocks": 174720, 00:15:02.338 "percent": 88 00:15:02.338 } 00:15:02.338 }, 00:15:02.338 "base_bdevs_list": [ 00:15:02.338 { 00:15:02.338 "name": "spare", 00:15:02.338 "uuid": "0af1829b-866f-5890-a8a2-14229f30a473", 00:15:02.338 "is_configured": true, 00:15:02.338 "data_offset": 0, 00:15:02.338 "data_size": 65536 00:15:02.338 }, 00:15:02.338 { 00:15:02.338 "name": "BaseBdev2", 00:15:02.338 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:15:02.338 "is_configured": true, 00:15:02.338 "data_offset": 0, 00:15:02.338 "data_size": 65536 00:15:02.338 }, 00:15:02.338 { 00:15:02.338 "name": "BaseBdev3", 00:15:02.338 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:15:02.338 "is_configured": true, 00:15:02.338 "data_offset": 0, 00:15:02.338 "data_size": 65536 00:15:02.338 }, 00:15:02.338 { 00:15:02.338 "name": "BaseBdev4", 00:15:02.338 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:15:02.338 "is_configured": true, 00:15:02.338 "data_offset": 0, 00:15:02.338 "data_size": 65536 00:15:02.338 } 00:15:02.338 ] 00:15:02.338 }' 00:15:02.338 19:05:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:02.338 19:05:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:02.338 19:05:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:02.338 19:05:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:02.338 19:05:19 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:03.275 [2024-12-05 19:05:20.773428] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:03.275 [2024-12-05 19:05:20.773570] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:03.275 [2024-12-05 19:05:20.773633] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:03.275 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:03.275 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:03.275 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:03.275 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:03.275 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:03.275 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:03.275 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.275 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:03.275 19:05:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.275 19:05:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.275 19:05:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.535 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:03.535 "name": "raid_bdev1", 00:15:03.535 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:15:03.535 "strip_size_kb": 64, 00:15:03.535 "state": "online", 00:15:03.535 "raid_level": "raid5f", 00:15:03.535 "superblock": false, 00:15:03.535 "num_base_bdevs": 4, 00:15:03.535 "num_base_bdevs_discovered": 4, 00:15:03.535 "num_base_bdevs_operational": 4, 00:15:03.535 "base_bdevs_list": [ 00:15:03.535 { 00:15:03.535 "name": "spare", 00:15:03.535 "uuid": "0af1829b-866f-5890-a8a2-14229f30a473", 00:15:03.535 "is_configured": true, 00:15:03.535 "data_offset": 0, 00:15:03.535 "data_size": 65536 00:15:03.535 }, 00:15:03.535 { 00:15:03.535 "name": "BaseBdev2", 00:15:03.535 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:15:03.535 "is_configured": true, 00:15:03.535 "data_offset": 0, 00:15:03.535 "data_size": 65536 00:15:03.535 }, 00:15:03.535 { 00:15:03.535 "name": "BaseBdev3", 00:15:03.535 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:15:03.535 "is_configured": true, 00:15:03.535 "data_offset": 0, 00:15:03.535 "data_size": 65536 00:15:03.535 }, 00:15:03.535 { 00:15:03.535 "name": "BaseBdev4", 00:15:03.535 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:15:03.535 "is_configured": true, 00:15:03.535 "data_offset": 0, 00:15:03.535 "data_size": 65536 00:15:03.535 } 00:15:03.535 ] 00:15:03.535 }' 00:15:03.535 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:03.536 "name": "raid_bdev1", 00:15:03.536 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:15:03.536 "strip_size_kb": 64, 00:15:03.536 "state": "online", 00:15:03.536 "raid_level": "raid5f", 00:15:03.536 "superblock": false, 00:15:03.536 "num_base_bdevs": 4, 00:15:03.536 "num_base_bdevs_discovered": 4, 00:15:03.536 "num_base_bdevs_operational": 4, 00:15:03.536 "base_bdevs_list": [ 00:15:03.536 { 00:15:03.536 "name": "spare", 00:15:03.536 "uuid": "0af1829b-866f-5890-a8a2-14229f30a473", 00:15:03.536 "is_configured": true, 00:15:03.536 "data_offset": 0, 00:15:03.536 "data_size": 65536 00:15:03.536 }, 00:15:03.536 { 00:15:03.536 "name": "BaseBdev2", 00:15:03.536 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:15:03.536 "is_configured": true, 00:15:03.536 "data_offset": 0, 00:15:03.536 "data_size": 65536 00:15:03.536 }, 00:15:03.536 { 00:15:03.536 "name": "BaseBdev3", 00:15:03.536 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:15:03.536 "is_configured": true, 00:15:03.536 "data_offset": 0, 00:15:03.536 "data_size": 65536 00:15:03.536 }, 00:15:03.536 { 00:15:03.536 "name": "BaseBdev4", 00:15:03.536 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:15:03.536 "is_configured": true, 00:15:03.536 "data_offset": 0, 00:15:03.536 "data_size": 65536 00:15:03.536 } 00:15:03.536 ] 00:15:03.536 }' 00:15:03.536 19:05:20 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:03.536 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.796 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:03.796 "name": "raid_bdev1", 00:15:03.796 "uuid": "c0e61b82-bbcc-47f4-9267-226a99610529", 00:15:03.796 "strip_size_kb": 64, 00:15:03.796 "state": "online", 00:15:03.796 "raid_level": "raid5f", 00:15:03.796 "superblock": false, 00:15:03.796 "num_base_bdevs": 4, 00:15:03.796 "num_base_bdevs_discovered": 4, 00:15:03.796 "num_base_bdevs_operational": 4, 00:15:03.796 "base_bdevs_list": [ 00:15:03.796 { 00:15:03.796 "name": "spare", 00:15:03.796 "uuid": "0af1829b-866f-5890-a8a2-14229f30a473", 00:15:03.796 "is_configured": true, 00:15:03.797 "data_offset": 0, 00:15:03.797 "data_size": 65536 00:15:03.797 }, 00:15:03.797 { 00:15:03.797 "name": "BaseBdev2", 00:15:03.797 "uuid": "c68b73b9-d74d-5f37-beb5-f243abf9f1fc", 00:15:03.797 "is_configured": true, 00:15:03.797 "data_offset": 0, 00:15:03.797 "data_size": 65536 00:15:03.797 }, 00:15:03.797 { 00:15:03.797 "name": "BaseBdev3", 00:15:03.797 "uuid": "27c6a6b6-b500-5b6f-9202-93fdeab752d1", 00:15:03.797 "is_configured": true, 00:15:03.797 "data_offset": 0, 00:15:03.797 "data_size": 65536 00:15:03.797 }, 00:15:03.797 { 00:15:03.797 "name": "BaseBdev4", 00:15:03.797 "uuid": "84ef9ab2-268d-58bd-92c9-cee2c8b0094a", 00:15:03.797 "is_configured": true, 00:15:03.797 "data_offset": 0, 00:15:03.797 "data_size": 65536 00:15:03.797 } 00:15:03.797 ] 00:15:03.797 }' 00:15:03.797 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:03.797 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.057 [2024-12-05 19:05:21.517730] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:04.057 [2024-12-05 19:05:21.517807] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:04.057 [2024-12-05 19:05:21.517899] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:04.057 [2024-12-05 19:05:21.518009] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:04.057 [2024-12-05 19:05:21.518071] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:04.057 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:04.317 /dev/nbd0 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:04.317 1+0 records in 00:15:04.317 1+0 records out 00:15:04.317 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000437537 s, 9.4 MB/s 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:04.317 19:05:21 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:04.576 /dev/nbd1 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # local i 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@877 -- # break 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:04.576 1+0 records in 00:15:04.576 1+0 records out 00:15:04.576 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000230372 s, 17.8 MB/s 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@890 -- # size=4096 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@893 -- # return 0 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:04.576 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:04.836 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 92470 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # '[' -z 92470 ']' 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@958 -- # kill -0 92470 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # uname 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 92470 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@972 -- # echo 'killing process with pid 92470' 00:15:05.096 killing process with pid 92470 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@973 -- # kill 92470 00:15:05.096 Received shutdown signal, test time was about 60.000000 seconds 00:15:05.096 00:15:05.096 Latency(us) 00:15:05.096 [2024-12-05T19:05:22.655Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:05.096 [2024-12-05T19:05:22.655Z] =================================================================================================================== 00:15:05.096 [2024-12-05T19:05:22.655Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:05.096 [2024-12-05 19:05:22.623502] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:05.096 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@978 -- # wait 92470 00:15:05.357 [2024-12-05 19:05:22.672473] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:05.357 19:05:22 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:15:05.357 ************************************ 00:15:05.357 END TEST raid5f_rebuild_test 00:15:05.357 ************************************ 00:15:05.357 00:15:05.357 real 0m17.125s 00:15:05.357 user 0m20.812s 00:15:05.357 sys 0m2.188s 00:15:05.357 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:05.357 19:05:22 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:05.617 19:05:22 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 4 true false true 00:15:05.617 19:05:22 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:15:05.617 19:05:22 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:05.617 19:05:22 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:05.617 ************************************ 00:15:05.617 START TEST raid5f_rebuild_test_sb 00:15:05.617 ************************************ 00:15:05.617 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid5f 4 true false true 00:15:05.617 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:15:05.617 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:15:05.617 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:05.617 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:05.617 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:05.617 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:05.617 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=92849 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 92849 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # '[' -z 92849 ']' 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:05.618 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:05.618 19:05:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.618 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:05.618 Zero copy mechanism will not be used. 00:15:05.618 [2024-12-05 19:05:23.065651] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:15:05.618 [2024-12-05 19:05:23.065776] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92849 ] 00:15:05.878 [2024-12-05 19:05:23.220725] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:05.878 [2024-12-05 19:05:23.245402] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:05.878 [2024-12-05 19:05:23.288824] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:05.878 [2024-12-05 19:05:23.288862] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # return 0 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.448 BaseBdev1_malloc 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.448 [2024-12-05 19:05:23.900857] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:06.448 [2024-12-05 19:05:23.900919] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:06.448 [2024-12-05 19:05:23.900951] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:06.448 [2024-12-05 19:05:23.900963] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:06.448 [2024-12-05 19:05:23.902985] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:06.448 [2024-12-05 19:05:23.903022] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:06.448 BaseBdev1 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.448 BaseBdev2_malloc 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.448 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.448 [2024-12-05 19:05:23.929368] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:06.448 [2024-12-05 19:05:23.929506] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:06.448 [2024-12-05 19:05:23.929545] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:06.448 [2024-12-05 19:05:23.929572] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:06.448 [2024-12-05 19:05:23.931564] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:06.449 [2024-12-05 19:05:23.931640] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:06.449 BaseBdev2 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.449 BaseBdev3_malloc 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.449 [2024-12-05 19:05:23.957890] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:15:06.449 [2024-12-05 19:05:23.957943] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:06.449 [2024-12-05 19:05:23.957964] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:06.449 [2024-12-05 19:05:23.957972] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:06.449 [2024-12-05 19:05:23.959935] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:06.449 [2024-12-05 19:05:23.960032] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:15:06.449 BaseBdev3 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.449 BaseBdev4_malloc 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.449 19:05:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.449 [2024-12-05 19:05:24.002168] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:15:06.449 [2024-12-05 19:05:24.002261] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:06.449 [2024-12-05 19:05:24.002308] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:06.449 [2024-12-05 19:05:24.002328] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:06.710 [2024-12-05 19:05:24.005939] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:06.710 [2024-12-05 19:05:24.005988] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:15:06.710 BaseBdev4 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.710 spare_malloc 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.710 spare_delay 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.710 [2024-12-05 19:05:24.043640] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:06.710 [2024-12-05 19:05:24.043701] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:06.710 [2024-12-05 19:05:24.043718] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:15:06.710 [2024-12-05 19:05:24.043726] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:06.710 [2024-12-05 19:05:24.045700] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:06.710 [2024-12-05 19:05:24.045734] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:06.710 spare 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.710 [2024-12-05 19:05:24.055712] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:06.710 [2024-12-05 19:05:24.057410] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:06.710 [2024-12-05 19:05:24.057474] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:06.710 [2024-12-05 19:05:24.057522] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:06.710 [2024-12-05 19:05:24.057707] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:06.710 [2024-12-05 19:05:24.057724] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:06.710 [2024-12-05 19:05:24.057945] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:06.710 [2024-12-05 19:05:24.058372] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:06.710 [2024-12-05 19:05:24.058390] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:06.710 [2024-12-05 19:05:24.058500] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:06.710 "name": "raid_bdev1", 00:15:06.710 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:06.710 "strip_size_kb": 64, 00:15:06.710 "state": "online", 00:15:06.710 "raid_level": "raid5f", 00:15:06.710 "superblock": true, 00:15:06.710 "num_base_bdevs": 4, 00:15:06.710 "num_base_bdevs_discovered": 4, 00:15:06.710 "num_base_bdevs_operational": 4, 00:15:06.710 "base_bdevs_list": [ 00:15:06.710 { 00:15:06.710 "name": "BaseBdev1", 00:15:06.710 "uuid": "24e127eb-b495-56f5-a726-1f56d7d7123f", 00:15:06.710 "is_configured": true, 00:15:06.710 "data_offset": 2048, 00:15:06.710 "data_size": 63488 00:15:06.710 }, 00:15:06.710 { 00:15:06.710 "name": "BaseBdev2", 00:15:06.710 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:06.710 "is_configured": true, 00:15:06.710 "data_offset": 2048, 00:15:06.710 "data_size": 63488 00:15:06.710 }, 00:15:06.710 { 00:15:06.710 "name": "BaseBdev3", 00:15:06.710 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:06.710 "is_configured": true, 00:15:06.710 "data_offset": 2048, 00:15:06.710 "data_size": 63488 00:15:06.710 }, 00:15:06.710 { 00:15:06.710 "name": "BaseBdev4", 00:15:06.710 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:06.710 "is_configured": true, 00:15:06.710 "data_offset": 2048, 00:15:06.710 "data_size": 63488 00:15:06.710 } 00:15:06.710 ] 00:15:06.710 }' 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:06.710 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.970 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:06.970 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:06.970 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:06.970 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.970 [2024-12-05 19:05:24.515652] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=190464 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:07.230 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:07.230 [2024-12-05 19:05:24.767057] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:15:07.230 /dev/nbd0 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:07.490 1+0 records in 00:15:07.490 1+0 records out 00:15:07.490 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000533249 s, 7.7 MB/s 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 192 00:15:07.490 19:05:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=496 oflag=direct 00:15:08.058 496+0 records in 00:15:08.058 496+0 records out 00:15:08.058 97517568 bytes (98 MB, 93 MiB) copied, 0.546096 s, 179 MB/s 00:15:08.058 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:08.058 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:08.058 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:08.058 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:08.058 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:15:08.058 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:08.058 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:08.058 [2024-12-05 19:05:25.608751] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:08.317 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:08.317 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:08.317 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:08.317 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:08.317 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:08.317 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.318 [2024-12-05 19:05:25.638342] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:08.318 "name": "raid_bdev1", 00:15:08.318 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:08.318 "strip_size_kb": 64, 00:15:08.318 "state": "online", 00:15:08.318 "raid_level": "raid5f", 00:15:08.318 "superblock": true, 00:15:08.318 "num_base_bdevs": 4, 00:15:08.318 "num_base_bdevs_discovered": 3, 00:15:08.318 "num_base_bdevs_operational": 3, 00:15:08.318 "base_bdevs_list": [ 00:15:08.318 { 00:15:08.318 "name": null, 00:15:08.318 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:08.318 "is_configured": false, 00:15:08.318 "data_offset": 0, 00:15:08.318 "data_size": 63488 00:15:08.318 }, 00:15:08.318 { 00:15:08.318 "name": "BaseBdev2", 00:15:08.318 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:08.318 "is_configured": true, 00:15:08.318 "data_offset": 2048, 00:15:08.318 "data_size": 63488 00:15:08.318 }, 00:15:08.318 { 00:15:08.318 "name": "BaseBdev3", 00:15:08.318 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:08.318 "is_configured": true, 00:15:08.318 "data_offset": 2048, 00:15:08.318 "data_size": 63488 00:15:08.318 }, 00:15:08.318 { 00:15:08.318 "name": "BaseBdev4", 00:15:08.318 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:08.318 "is_configured": true, 00:15:08.318 "data_offset": 2048, 00:15:08.318 "data_size": 63488 00:15:08.318 } 00:15:08.318 ] 00:15:08.318 }' 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:08.318 19:05:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.577 19:05:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:08.577 19:05:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:08.577 19:05:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.577 [2024-12-05 19:05:26.117513] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:08.577 [2024-12-05 19:05:26.121856] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000270a0 00:15:08.577 19:05:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:08.577 19:05:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:08.577 [2024-12-05 19:05:26.124125] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:09.961 "name": "raid_bdev1", 00:15:09.961 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:09.961 "strip_size_kb": 64, 00:15:09.961 "state": "online", 00:15:09.961 "raid_level": "raid5f", 00:15:09.961 "superblock": true, 00:15:09.961 "num_base_bdevs": 4, 00:15:09.961 "num_base_bdevs_discovered": 4, 00:15:09.961 "num_base_bdevs_operational": 4, 00:15:09.961 "process": { 00:15:09.961 "type": "rebuild", 00:15:09.961 "target": "spare", 00:15:09.961 "progress": { 00:15:09.961 "blocks": 19200, 00:15:09.961 "percent": 10 00:15:09.961 } 00:15:09.961 }, 00:15:09.961 "base_bdevs_list": [ 00:15:09.961 { 00:15:09.961 "name": "spare", 00:15:09.961 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:09.961 "is_configured": true, 00:15:09.961 "data_offset": 2048, 00:15:09.961 "data_size": 63488 00:15:09.961 }, 00:15:09.961 { 00:15:09.961 "name": "BaseBdev2", 00:15:09.961 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:09.961 "is_configured": true, 00:15:09.961 "data_offset": 2048, 00:15:09.961 "data_size": 63488 00:15:09.961 }, 00:15:09.961 { 00:15:09.961 "name": "BaseBdev3", 00:15:09.961 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:09.961 "is_configured": true, 00:15:09.961 "data_offset": 2048, 00:15:09.961 "data_size": 63488 00:15:09.961 }, 00:15:09.961 { 00:15:09.961 "name": "BaseBdev4", 00:15:09.961 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:09.961 "is_configured": true, 00:15:09.961 "data_offset": 2048, 00:15:09.961 "data_size": 63488 00:15:09.961 } 00:15:09.961 ] 00:15:09.961 }' 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.961 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:09.961 [2024-12-05 19:05:27.272889] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:09.961 [2024-12-05 19:05:27.329473] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:09.961 [2024-12-05 19:05:27.329534] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:09.961 [2024-12-05 19:05:27.329552] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:09.962 [2024-12-05 19:05:27.329559] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:09.962 "name": "raid_bdev1", 00:15:09.962 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:09.962 "strip_size_kb": 64, 00:15:09.962 "state": "online", 00:15:09.962 "raid_level": "raid5f", 00:15:09.962 "superblock": true, 00:15:09.962 "num_base_bdevs": 4, 00:15:09.962 "num_base_bdevs_discovered": 3, 00:15:09.962 "num_base_bdevs_operational": 3, 00:15:09.962 "base_bdevs_list": [ 00:15:09.962 { 00:15:09.962 "name": null, 00:15:09.962 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:09.962 "is_configured": false, 00:15:09.962 "data_offset": 0, 00:15:09.962 "data_size": 63488 00:15:09.962 }, 00:15:09.962 { 00:15:09.962 "name": "BaseBdev2", 00:15:09.962 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:09.962 "is_configured": true, 00:15:09.962 "data_offset": 2048, 00:15:09.962 "data_size": 63488 00:15:09.962 }, 00:15:09.962 { 00:15:09.962 "name": "BaseBdev3", 00:15:09.962 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:09.962 "is_configured": true, 00:15:09.962 "data_offset": 2048, 00:15:09.962 "data_size": 63488 00:15:09.962 }, 00:15:09.962 { 00:15:09.962 "name": "BaseBdev4", 00:15:09.962 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:09.962 "is_configured": true, 00:15:09.962 "data_offset": 2048, 00:15:09.962 "data_size": 63488 00:15:09.962 } 00:15:09.962 ] 00:15:09.962 }' 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:09.962 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:10.222 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:10.222 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:10.222 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:10.222 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:10.222 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:10.222 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:10.222 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:10.222 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.483 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:10.483 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.483 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:10.483 "name": "raid_bdev1", 00:15:10.483 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:10.483 "strip_size_kb": 64, 00:15:10.483 "state": "online", 00:15:10.483 "raid_level": "raid5f", 00:15:10.483 "superblock": true, 00:15:10.483 "num_base_bdevs": 4, 00:15:10.483 "num_base_bdevs_discovered": 3, 00:15:10.483 "num_base_bdevs_operational": 3, 00:15:10.483 "base_bdevs_list": [ 00:15:10.483 { 00:15:10.483 "name": null, 00:15:10.483 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:10.483 "is_configured": false, 00:15:10.483 "data_offset": 0, 00:15:10.483 "data_size": 63488 00:15:10.483 }, 00:15:10.483 { 00:15:10.483 "name": "BaseBdev2", 00:15:10.483 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:10.483 "is_configured": true, 00:15:10.483 "data_offset": 2048, 00:15:10.483 "data_size": 63488 00:15:10.483 }, 00:15:10.483 { 00:15:10.483 "name": "BaseBdev3", 00:15:10.483 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:10.483 "is_configured": true, 00:15:10.483 "data_offset": 2048, 00:15:10.483 "data_size": 63488 00:15:10.483 }, 00:15:10.483 { 00:15:10.483 "name": "BaseBdev4", 00:15:10.483 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:10.483 "is_configured": true, 00:15:10.483 "data_offset": 2048, 00:15:10.483 "data_size": 63488 00:15:10.483 } 00:15:10.483 ] 00:15:10.483 }' 00:15:10.483 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:10.483 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:10.483 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:10.483 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:10.483 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:10.483 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.483 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:10.483 [2024-12-05 19:05:27.906012] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:10.483 [2024-12-05 19:05:27.909475] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027170 00:15:10.483 [2024-12-05 19:05:27.911660] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:10.483 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.483 19:05:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:11.424 19:05:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:11.424 19:05:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:11.424 19:05:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:11.424 19:05:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:11.424 19:05:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:11.424 19:05:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.424 19:05:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:11.424 19:05:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.424 19:05:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:11.424 19:05:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.424 19:05:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:11.424 "name": "raid_bdev1", 00:15:11.424 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:11.424 "strip_size_kb": 64, 00:15:11.424 "state": "online", 00:15:11.424 "raid_level": "raid5f", 00:15:11.424 "superblock": true, 00:15:11.424 "num_base_bdevs": 4, 00:15:11.424 "num_base_bdevs_discovered": 4, 00:15:11.424 "num_base_bdevs_operational": 4, 00:15:11.424 "process": { 00:15:11.424 "type": "rebuild", 00:15:11.424 "target": "spare", 00:15:11.424 "progress": { 00:15:11.424 "blocks": 19200, 00:15:11.424 "percent": 10 00:15:11.424 } 00:15:11.424 }, 00:15:11.424 "base_bdevs_list": [ 00:15:11.424 { 00:15:11.424 "name": "spare", 00:15:11.424 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:11.424 "is_configured": true, 00:15:11.424 "data_offset": 2048, 00:15:11.424 "data_size": 63488 00:15:11.424 }, 00:15:11.424 { 00:15:11.424 "name": "BaseBdev2", 00:15:11.424 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:11.424 "is_configured": true, 00:15:11.424 "data_offset": 2048, 00:15:11.424 "data_size": 63488 00:15:11.424 }, 00:15:11.424 { 00:15:11.424 "name": "BaseBdev3", 00:15:11.424 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:11.424 "is_configured": true, 00:15:11.424 "data_offset": 2048, 00:15:11.424 "data_size": 63488 00:15:11.424 }, 00:15:11.424 { 00:15:11.424 "name": "BaseBdev4", 00:15:11.424 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:11.424 "is_configured": true, 00:15:11.424 "data_offset": 2048, 00:15:11.424 "data_size": 63488 00:15:11.424 } 00:15:11.424 ] 00:15:11.424 }' 00:15:11.424 19:05:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:15:11.685 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=519 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:11.685 "name": "raid_bdev1", 00:15:11.685 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:11.685 "strip_size_kb": 64, 00:15:11.685 "state": "online", 00:15:11.685 "raid_level": "raid5f", 00:15:11.685 "superblock": true, 00:15:11.685 "num_base_bdevs": 4, 00:15:11.685 "num_base_bdevs_discovered": 4, 00:15:11.685 "num_base_bdevs_operational": 4, 00:15:11.685 "process": { 00:15:11.685 "type": "rebuild", 00:15:11.685 "target": "spare", 00:15:11.685 "progress": { 00:15:11.685 "blocks": 21120, 00:15:11.685 "percent": 11 00:15:11.685 } 00:15:11.685 }, 00:15:11.685 "base_bdevs_list": [ 00:15:11.685 { 00:15:11.685 "name": "spare", 00:15:11.685 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:11.685 "is_configured": true, 00:15:11.685 "data_offset": 2048, 00:15:11.685 "data_size": 63488 00:15:11.685 }, 00:15:11.685 { 00:15:11.685 "name": "BaseBdev2", 00:15:11.685 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:11.685 "is_configured": true, 00:15:11.685 "data_offset": 2048, 00:15:11.685 "data_size": 63488 00:15:11.685 }, 00:15:11.685 { 00:15:11.685 "name": "BaseBdev3", 00:15:11.685 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:11.685 "is_configured": true, 00:15:11.685 "data_offset": 2048, 00:15:11.685 "data_size": 63488 00:15:11.685 }, 00:15:11.685 { 00:15:11.685 "name": "BaseBdev4", 00:15:11.685 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:11.685 "is_configured": true, 00:15:11.685 "data_offset": 2048, 00:15:11.685 "data_size": 63488 00:15:11.685 } 00:15:11.685 ] 00:15:11.685 }' 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:11.685 19:05:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:13.066 "name": "raid_bdev1", 00:15:13.066 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:13.066 "strip_size_kb": 64, 00:15:13.066 "state": "online", 00:15:13.066 "raid_level": "raid5f", 00:15:13.066 "superblock": true, 00:15:13.066 "num_base_bdevs": 4, 00:15:13.066 "num_base_bdevs_discovered": 4, 00:15:13.066 "num_base_bdevs_operational": 4, 00:15:13.066 "process": { 00:15:13.066 "type": "rebuild", 00:15:13.066 "target": "spare", 00:15:13.066 "progress": { 00:15:13.066 "blocks": 42240, 00:15:13.066 "percent": 22 00:15:13.066 } 00:15:13.066 }, 00:15:13.066 "base_bdevs_list": [ 00:15:13.066 { 00:15:13.066 "name": "spare", 00:15:13.066 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:13.066 "is_configured": true, 00:15:13.066 "data_offset": 2048, 00:15:13.066 "data_size": 63488 00:15:13.066 }, 00:15:13.066 { 00:15:13.066 "name": "BaseBdev2", 00:15:13.066 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:13.066 "is_configured": true, 00:15:13.066 "data_offset": 2048, 00:15:13.066 "data_size": 63488 00:15:13.066 }, 00:15:13.066 { 00:15:13.066 "name": "BaseBdev3", 00:15:13.066 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:13.066 "is_configured": true, 00:15:13.066 "data_offset": 2048, 00:15:13.066 "data_size": 63488 00:15:13.066 }, 00:15:13.066 { 00:15:13.066 "name": "BaseBdev4", 00:15:13.066 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:13.066 "is_configured": true, 00:15:13.066 "data_offset": 2048, 00:15:13.066 "data_size": 63488 00:15:13.066 } 00:15:13.066 ] 00:15:13.066 }' 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:13.066 19:05:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:14.006 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:14.006 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:14.006 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:14.006 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:14.006 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:14.006 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:14.006 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:14.006 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:14.006 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:14.006 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:14.006 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:14.006 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:14.006 "name": "raid_bdev1", 00:15:14.007 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:14.007 "strip_size_kb": 64, 00:15:14.007 "state": "online", 00:15:14.007 "raid_level": "raid5f", 00:15:14.007 "superblock": true, 00:15:14.007 "num_base_bdevs": 4, 00:15:14.007 "num_base_bdevs_discovered": 4, 00:15:14.007 "num_base_bdevs_operational": 4, 00:15:14.007 "process": { 00:15:14.007 "type": "rebuild", 00:15:14.007 "target": "spare", 00:15:14.007 "progress": { 00:15:14.007 "blocks": 65280, 00:15:14.007 "percent": 34 00:15:14.007 } 00:15:14.007 }, 00:15:14.007 "base_bdevs_list": [ 00:15:14.007 { 00:15:14.007 "name": "spare", 00:15:14.007 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:14.007 "is_configured": true, 00:15:14.007 "data_offset": 2048, 00:15:14.007 "data_size": 63488 00:15:14.007 }, 00:15:14.007 { 00:15:14.007 "name": "BaseBdev2", 00:15:14.007 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:14.007 "is_configured": true, 00:15:14.007 "data_offset": 2048, 00:15:14.007 "data_size": 63488 00:15:14.007 }, 00:15:14.007 { 00:15:14.007 "name": "BaseBdev3", 00:15:14.007 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:14.007 "is_configured": true, 00:15:14.007 "data_offset": 2048, 00:15:14.007 "data_size": 63488 00:15:14.007 }, 00:15:14.007 { 00:15:14.007 "name": "BaseBdev4", 00:15:14.007 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:14.007 "is_configured": true, 00:15:14.007 "data_offset": 2048, 00:15:14.007 "data_size": 63488 00:15:14.007 } 00:15:14.007 ] 00:15:14.007 }' 00:15:14.007 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:14.007 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:14.007 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:14.007 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:14.007 19:05:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:14.948 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:14.948 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:14.948 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:14.948 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:14.948 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:14.948 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:15.208 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.208 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:15.208 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:15.208 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:15.208 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:15.208 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:15.208 "name": "raid_bdev1", 00:15:15.208 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:15.208 "strip_size_kb": 64, 00:15:15.208 "state": "online", 00:15:15.208 "raid_level": "raid5f", 00:15:15.208 "superblock": true, 00:15:15.208 "num_base_bdevs": 4, 00:15:15.208 "num_base_bdevs_discovered": 4, 00:15:15.208 "num_base_bdevs_operational": 4, 00:15:15.208 "process": { 00:15:15.208 "type": "rebuild", 00:15:15.208 "target": "spare", 00:15:15.208 "progress": { 00:15:15.208 "blocks": 86400, 00:15:15.208 "percent": 45 00:15:15.208 } 00:15:15.208 }, 00:15:15.208 "base_bdevs_list": [ 00:15:15.208 { 00:15:15.208 "name": "spare", 00:15:15.208 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:15.208 "is_configured": true, 00:15:15.208 "data_offset": 2048, 00:15:15.208 "data_size": 63488 00:15:15.208 }, 00:15:15.208 { 00:15:15.209 "name": "BaseBdev2", 00:15:15.209 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:15.209 "is_configured": true, 00:15:15.209 "data_offset": 2048, 00:15:15.209 "data_size": 63488 00:15:15.209 }, 00:15:15.209 { 00:15:15.209 "name": "BaseBdev3", 00:15:15.209 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:15.209 "is_configured": true, 00:15:15.209 "data_offset": 2048, 00:15:15.209 "data_size": 63488 00:15:15.209 }, 00:15:15.209 { 00:15:15.209 "name": "BaseBdev4", 00:15:15.209 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:15.209 "is_configured": true, 00:15:15.209 "data_offset": 2048, 00:15:15.209 "data_size": 63488 00:15:15.209 } 00:15:15.209 ] 00:15:15.209 }' 00:15:15.209 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:15.209 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:15.209 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:15.209 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:15.209 19:05:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:16.150 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:16.150 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:16.150 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:16.150 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:16.150 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:16.150 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:16.150 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:16.150 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.150 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:16.150 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.150 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.150 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:16.150 "name": "raid_bdev1", 00:15:16.150 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:16.150 "strip_size_kb": 64, 00:15:16.150 "state": "online", 00:15:16.150 "raid_level": "raid5f", 00:15:16.150 "superblock": true, 00:15:16.150 "num_base_bdevs": 4, 00:15:16.150 "num_base_bdevs_discovered": 4, 00:15:16.150 "num_base_bdevs_operational": 4, 00:15:16.150 "process": { 00:15:16.150 "type": "rebuild", 00:15:16.150 "target": "spare", 00:15:16.150 "progress": { 00:15:16.150 "blocks": 109440, 00:15:16.150 "percent": 57 00:15:16.150 } 00:15:16.150 }, 00:15:16.150 "base_bdevs_list": [ 00:15:16.150 { 00:15:16.150 "name": "spare", 00:15:16.150 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:16.150 "is_configured": true, 00:15:16.150 "data_offset": 2048, 00:15:16.150 "data_size": 63488 00:15:16.150 }, 00:15:16.150 { 00:15:16.150 "name": "BaseBdev2", 00:15:16.150 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:16.150 "is_configured": true, 00:15:16.150 "data_offset": 2048, 00:15:16.150 "data_size": 63488 00:15:16.150 }, 00:15:16.150 { 00:15:16.150 "name": "BaseBdev3", 00:15:16.150 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:16.150 "is_configured": true, 00:15:16.150 "data_offset": 2048, 00:15:16.150 "data_size": 63488 00:15:16.150 }, 00:15:16.150 { 00:15:16.150 "name": "BaseBdev4", 00:15:16.150 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:16.150 "is_configured": true, 00:15:16.150 "data_offset": 2048, 00:15:16.150 "data_size": 63488 00:15:16.150 } 00:15:16.150 ] 00:15:16.150 }' 00:15:16.150 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:16.411 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:16.411 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:16.411 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:16.411 19:05:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:17.352 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:17.352 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:17.352 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:17.352 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:17.352 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:17.352 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:17.352 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:17.352 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:17.352 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:17.352 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:17.352 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:17.352 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:17.352 "name": "raid_bdev1", 00:15:17.352 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:17.352 "strip_size_kb": 64, 00:15:17.352 "state": "online", 00:15:17.352 "raid_level": "raid5f", 00:15:17.352 "superblock": true, 00:15:17.352 "num_base_bdevs": 4, 00:15:17.352 "num_base_bdevs_discovered": 4, 00:15:17.352 "num_base_bdevs_operational": 4, 00:15:17.352 "process": { 00:15:17.352 "type": "rebuild", 00:15:17.352 "target": "spare", 00:15:17.352 "progress": { 00:15:17.352 "blocks": 130560, 00:15:17.352 "percent": 68 00:15:17.352 } 00:15:17.352 }, 00:15:17.352 "base_bdevs_list": [ 00:15:17.352 { 00:15:17.352 "name": "spare", 00:15:17.352 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:17.352 "is_configured": true, 00:15:17.352 "data_offset": 2048, 00:15:17.352 "data_size": 63488 00:15:17.352 }, 00:15:17.352 { 00:15:17.352 "name": "BaseBdev2", 00:15:17.352 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:17.352 "is_configured": true, 00:15:17.352 "data_offset": 2048, 00:15:17.352 "data_size": 63488 00:15:17.352 }, 00:15:17.352 { 00:15:17.352 "name": "BaseBdev3", 00:15:17.352 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:17.352 "is_configured": true, 00:15:17.352 "data_offset": 2048, 00:15:17.352 "data_size": 63488 00:15:17.352 }, 00:15:17.352 { 00:15:17.352 "name": "BaseBdev4", 00:15:17.352 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:17.352 "is_configured": true, 00:15:17.352 "data_offset": 2048, 00:15:17.352 "data_size": 63488 00:15:17.352 } 00:15:17.352 ] 00:15:17.352 }' 00:15:17.352 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:17.352 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:17.352 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:17.613 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:17.613 19:05:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:18.655 19:05:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:18.655 19:05:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:18.655 19:05:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:18.655 19:05:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:18.655 19:05:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:18.655 19:05:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:18.655 19:05:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:18.655 19:05:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:18.655 19:05:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:18.655 19:05:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:18.655 19:05:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:18.655 19:05:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:18.655 "name": "raid_bdev1", 00:15:18.655 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:18.655 "strip_size_kb": 64, 00:15:18.655 "state": "online", 00:15:18.655 "raid_level": "raid5f", 00:15:18.655 "superblock": true, 00:15:18.655 "num_base_bdevs": 4, 00:15:18.655 "num_base_bdevs_discovered": 4, 00:15:18.655 "num_base_bdevs_operational": 4, 00:15:18.655 "process": { 00:15:18.655 "type": "rebuild", 00:15:18.655 "target": "spare", 00:15:18.655 "progress": { 00:15:18.655 "blocks": 153600, 00:15:18.655 "percent": 80 00:15:18.655 } 00:15:18.655 }, 00:15:18.655 "base_bdevs_list": [ 00:15:18.655 { 00:15:18.655 "name": "spare", 00:15:18.655 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:18.655 "is_configured": true, 00:15:18.655 "data_offset": 2048, 00:15:18.655 "data_size": 63488 00:15:18.655 }, 00:15:18.655 { 00:15:18.655 "name": "BaseBdev2", 00:15:18.655 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:18.655 "is_configured": true, 00:15:18.655 "data_offset": 2048, 00:15:18.655 "data_size": 63488 00:15:18.655 }, 00:15:18.655 { 00:15:18.655 "name": "BaseBdev3", 00:15:18.655 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:18.655 "is_configured": true, 00:15:18.655 "data_offset": 2048, 00:15:18.655 "data_size": 63488 00:15:18.655 }, 00:15:18.655 { 00:15:18.655 "name": "BaseBdev4", 00:15:18.655 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:18.655 "is_configured": true, 00:15:18.655 "data_offset": 2048, 00:15:18.655 "data_size": 63488 00:15:18.655 } 00:15:18.655 ] 00:15:18.655 }' 00:15:18.655 19:05:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:18.655 19:05:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:18.655 19:05:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:18.655 19:05:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:18.655 19:05:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:19.596 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:19.596 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:19.596 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:19.596 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:19.596 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:19.596 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:19.596 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:19.596 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:19.596 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:19.596 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:19.596 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:19.596 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:19.596 "name": "raid_bdev1", 00:15:19.596 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:19.596 "strip_size_kb": 64, 00:15:19.596 "state": "online", 00:15:19.596 "raid_level": "raid5f", 00:15:19.596 "superblock": true, 00:15:19.596 "num_base_bdevs": 4, 00:15:19.596 "num_base_bdevs_discovered": 4, 00:15:19.596 "num_base_bdevs_operational": 4, 00:15:19.596 "process": { 00:15:19.596 "type": "rebuild", 00:15:19.596 "target": "spare", 00:15:19.596 "progress": { 00:15:19.596 "blocks": 174720, 00:15:19.596 "percent": 91 00:15:19.596 } 00:15:19.596 }, 00:15:19.596 "base_bdevs_list": [ 00:15:19.596 { 00:15:19.596 "name": "spare", 00:15:19.596 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:19.596 "is_configured": true, 00:15:19.596 "data_offset": 2048, 00:15:19.596 "data_size": 63488 00:15:19.596 }, 00:15:19.596 { 00:15:19.596 "name": "BaseBdev2", 00:15:19.596 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:19.596 "is_configured": true, 00:15:19.596 "data_offset": 2048, 00:15:19.596 "data_size": 63488 00:15:19.596 }, 00:15:19.596 { 00:15:19.596 "name": "BaseBdev3", 00:15:19.596 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:19.596 "is_configured": true, 00:15:19.596 "data_offset": 2048, 00:15:19.596 "data_size": 63488 00:15:19.596 }, 00:15:19.596 { 00:15:19.596 "name": "BaseBdev4", 00:15:19.596 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:19.596 "is_configured": true, 00:15:19.596 "data_offset": 2048, 00:15:19.596 "data_size": 63488 00:15:19.596 } 00:15:19.596 ] 00:15:19.596 }' 00:15:19.596 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:19.856 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:19.856 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:19.856 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:19.856 19:05:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:20.425 [2024-12-05 19:05:37.950090] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:20.425 [2024-12-05 19:05:37.950211] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:20.425 [2024-12-05 19:05:37.950371] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:20.685 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:20.685 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:20.685 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:20.685 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:20.685 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:20.685 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:20.685 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:20.685 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:20.685 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:20.685 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:20.946 "name": "raid_bdev1", 00:15:20.946 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:20.946 "strip_size_kb": 64, 00:15:20.946 "state": "online", 00:15:20.946 "raid_level": "raid5f", 00:15:20.946 "superblock": true, 00:15:20.946 "num_base_bdevs": 4, 00:15:20.946 "num_base_bdevs_discovered": 4, 00:15:20.946 "num_base_bdevs_operational": 4, 00:15:20.946 "base_bdevs_list": [ 00:15:20.946 { 00:15:20.946 "name": "spare", 00:15:20.946 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:20.946 "is_configured": true, 00:15:20.946 "data_offset": 2048, 00:15:20.946 "data_size": 63488 00:15:20.946 }, 00:15:20.946 { 00:15:20.946 "name": "BaseBdev2", 00:15:20.946 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:20.946 "is_configured": true, 00:15:20.946 "data_offset": 2048, 00:15:20.946 "data_size": 63488 00:15:20.946 }, 00:15:20.946 { 00:15:20.946 "name": "BaseBdev3", 00:15:20.946 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:20.946 "is_configured": true, 00:15:20.946 "data_offset": 2048, 00:15:20.946 "data_size": 63488 00:15:20.946 }, 00:15:20.946 { 00:15:20.946 "name": "BaseBdev4", 00:15:20.946 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:20.946 "is_configured": true, 00:15:20.946 "data_offset": 2048, 00:15:20.946 "data_size": 63488 00:15:20.946 } 00:15:20.946 ] 00:15:20.946 }' 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:20.946 "name": "raid_bdev1", 00:15:20.946 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:20.946 "strip_size_kb": 64, 00:15:20.946 "state": "online", 00:15:20.946 "raid_level": "raid5f", 00:15:20.946 "superblock": true, 00:15:20.946 "num_base_bdevs": 4, 00:15:20.946 "num_base_bdevs_discovered": 4, 00:15:20.946 "num_base_bdevs_operational": 4, 00:15:20.946 "base_bdevs_list": [ 00:15:20.946 { 00:15:20.946 "name": "spare", 00:15:20.946 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:20.946 "is_configured": true, 00:15:20.946 "data_offset": 2048, 00:15:20.946 "data_size": 63488 00:15:20.946 }, 00:15:20.946 { 00:15:20.946 "name": "BaseBdev2", 00:15:20.946 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:20.946 "is_configured": true, 00:15:20.946 "data_offset": 2048, 00:15:20.946 "data_size": 63488 00:15:20.946 }, 00:15:20.946 { 00:15:20.946 "name": "BaseBdev3", 00:15:20.946 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:20.946 "is_configured": true, 00:15:20.946 "data_offset": 2048, 00:15:20.946 "data_size": 63488 00:15:20.946 }, 00:15:20.946 { 00:15:20.946 "name": "BaseBdev4", 00:15:20.946 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:20.946 "is_configured": true, 00:15:20.946 "data_offset": 2048, 00:15:20.946 "data_size": 63488 00:15:20.946 } 00:15:20.946 ] 00:15:20.946 }' 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:20.946 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:21.206 "name": "raid_bdev1", 00:15:21.206 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:21.206 "strip_size_kb": 64, 00:15:21.206 "state": "online", 00:15:21.206 "raid_level": "raid5f", 00:15:21.206 "superblock": true, 00:15:21.206 "num_base_bdevs": 4, 00:15:21.206 "num_base_bdevs_discovered": 4, 00:15:21.206 "num_base_bdevs_operational": 4, 00:15:21.206 "base_bdevs_list": [ 00:15:21.206 { 00:15:21.206 "name": "spare", 00:15:21.206 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:21.206 "is_configured": true, 00:15:21.206 "data_offset": 2048, 00:15:21.206 "data_size": 63488 00:15:21.206 }, 00:15:21.206 { 00:15:21.206 "name": "BaseBdev2", 00:15:21.206 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:21.206 "is_configured": true, 00:15:21.206 "data_offset": 2048, 00:15:21.206 "data_size": 63488 00:15:21.206 }, 00:15:21.206 { 00:15:21.206 "name": "BaseBdev3", 00:15:21.206 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:21.206 "is_configured": true, 00:15:21.206 "data_offset": 2048, 00:15:21.206 "data_size": 63488 00:15:21.206 }, 00:15:21.206 { 00:15:21.206 "name": "BaseBdev4", 00:15:21.206 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:21.206 "is_configured": true, 00:15:21.206 "data_offset": 2048, 00:15:21.206 "data_size": 63488 00:15:21.206 } 00:15:21.206 ] 00:15:21.206 }' 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:21.206 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.467 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:21.467 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.467 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.467 [2024-12-05 19:05:38.965951] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:21.467 [2024-12-05 19:05:38.966030] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:21.467 [2024-12-05 19:05:38.966132] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:21.467 [2024-12-05 19:05:38.966247] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:21.467 [2024-12-05 19:05:38.966336] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:21.467 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.467 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.467 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:21.467 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:21.467 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:15:21.467 19:05:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:21.467 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:21.467 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:21.467 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:21.467 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:21.467 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:21.467 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:21.467 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:21.467 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:21.467 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:21.467 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:15:21.467 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:21.467 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:21.467 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:21.729 /dev/nbd0 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:21.729 1+0 records in 00:15:21.729 1+0 records out 00:15:21.729 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000424255 s, 9.7 MB/s 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:21.729 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:21.989 /dev/nbd1 00:15:21.989 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:21.989 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:21.989 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:15:21.989 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # local i 00:15:21.989 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:21.989 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:21.989 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:15:21.989 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@877 -- # break 00:15:21.989 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:21.989 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:21.989 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:21.989 1+0 records in 00:15:21.989 1+0 records out 00:15:21.989 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000387215 s, 10.6 MB/s 00:15:21.989 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:21.990 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@890 -- # size=4096 00:15:21.990 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:21.990 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:21.990 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@893 -- # return 0 00:15:21.990 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:21.990 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:21.990 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:22.250 19:05:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:22.512 [2024-12-05 19:05:40.056049] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:22.512 [2024-12-05 19:05:40.056111] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:22.512 [2024-12-05 19:05:40.056132] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:15:22.512 [2024-12-05 19:05:40.056142] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:22.512 [2024-12-05 19:05:40.058513] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:22.512 [2024-12-05 19:05:40.058555] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:22.512 [2024-12-05 19:05:40.058624] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:22.512 [2024-12-05 19:05:40.058661] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:22.512 [2024-12-05 19:05:40.058784] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:22.512 [2024-12-05 19:05:40.058888] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:22.512 [2024-12-05 19:05:40.058947] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:22.512 spare 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:22.512 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:22.772 [2024-12-05 19:05:40.158839] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:15:22.772 [2024-12-05 19:05:40.158861] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:22.772 [2024-12-05 19:05:40.159111] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000045820 00:15:22.772 [2024-12-05 19:05:40.159528] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:15:22.772 [2024-12-05 19:05:40.159542] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:15:22.772 [2024-12-05 19:05:40.159654] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:22.772 "name": "raid_bdev1", 00:15:22.772 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:22.772 "strip_size_kb": 64, 00:15:22.772 "state": "online", 00:15:22.772 "raid_level": "raid5f", 00:15:22.772 "superblock": true, 00:15:22.772 "num_base_bdevs": 4, 00:15:22.772 "num_base_bdevs_discovered": 4, 00:15:22.772 "num_base_bdevs_operational": 4, 00:15:22.772 "base_bdevs_list": [ 00:15:22.772 { 00:15:22.772 "name": "spare", 00:15:22.772 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:22.772 "is_configured": true, 00:15:22.772 "data_offset": 2048, 00:15:22.772 "data_size": 63488 00:15:22.772 }, 00:15:22.772 { 00:15:22.772 "name": "BaseBdev2", 00:15:22.772 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:22.772 "is_configured": true, 00:15:22.772 "data_offset": 2048, 00:15:22.772 "data_size": 63488 00:15:22.772 }, 00:15:22.772 { 00:15:22.772 "name": "BaseBdev3", 00:15:22.772 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:22.772 "is_configured": true, 00:15:22.772 "data_offset": 2048, 00:15:22.772 "data_size": 63488 00:15:22.772 }, 00:15:22.772 { 00:15:22.772 "name": "BaseBdev4", 00:15:22.772 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:22.772 "is_configured": true, 00:15:22.772 "data_offset": 2048, 00:15:22.772 "data_size": 63488 00:15:22.772 } 00:15:22.772 ] 00:15:22.772 }' 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:22.772 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:23.343 "name": "raid_bdev1", 00:15:23.343 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:23.343 "strip_size_kb": 64, 00:15:23.343 "state": "online", 00:15:23.343 "raid_level": "raid5f", 00:15:23.343 "superblock": true, 00:15:23.343 "num_base_bdevs": 4, 00:15:23.343 "num_base_bdevs_discovered": 4, 00:15:23.343 "num_base_bdevs_operational": 4, 00:15:23.343 "base_bdevs_list": [ 00:15:23.343 { 00:15:23.343 "name": "spare", 00:15:23.343 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:23.343 "is_configured": true, 00:15:23.343 "data_offset": 2048, 00:15:23.343 "data_size": 63488 00:15:23.343 }, 00:15:23.343 { 00:15:23.343 "name": "BaseBdev2", 00:15:23.343 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:23.343 "is_configured": true, 00:15:23.343 "data_offset": 2048, 00:15:23.343 "data_size": 63488 00:15:23.343 }, 00:15:23.343 { 00:15:23.343 "name": "BaseBdev3", 00:15:23.343 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:23.343 "is_configured": true, 00:15:23.343 "data_offset": 2048, 00:15:23.343 "data_size": 63488 00:15:23.343 }, 00:15:23.343 { 00:15:23.343 "name": "BaseBdev4", 00:15:23.343 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:23.343 "is_configured": true, 00:15:23.343 "data_offset": 2048, 00:15:23.343 "data_size": 63488 00:15:23.343 } 00:15:23.343 ] 00:15:23.343 }' 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:23.343 [2024-12-05 19:05:40.852199] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:23.343 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:23.603 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:23.603 "name": "raid_bdev1", 00:15:23.603 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:23.603 "strip_size_kb": 64, 00:15:23.603 "state": "online", 00:15:23.603 "raid_level": "raid5f", 00:15:23.603 "superblock": true, 00:15:23.603 "num_base_bdevs": 4, 00:15:23.603 "num_base_bdevs_discovered": 3, 00:15:23.603 "num_base_bdevs_operational": 3, 00:15:23.603 "base_bdevs_list": [ 00:15:23.603 { 00:15:23.603 "name": null, 00:15:23.603 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:23.603 "is_configured": false, 00:15:23.603 "data_offset": 0, 00:15:23.603 "data_size": 63488 00:15:23.603 }, 00:15:23.603 { 00:15:23.603 "name": "BaseBdev2", 00:15:23.603 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:23.603 "is_configured": true, 00:15:23.603 "data_offset": 2048, 00:15:23.603 "data_size": 63488 00:15:23.603 }, 00:15:23.603 { 00:15:23.603 "name": "BaseBdev3", 00:15:23.603 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:23.603 "is_configured": true, 00:15:23.603 "data_offset": 2048, 00:15:23.603 "data_size": 63488 00:15:23.603 }, 00:15:23.603 { 00:15:23.603 "name": "BaseBdev4", 00:15:23.603 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:23.603 "is_configured": true, 00:15:23.603 "data_offset": 2048, 00:15:23.603 "data_size": 63488 00:15:23.603 } 00:15:23.603 ] 00:15:23.603 }' 00:15:23.603 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:23.603 19:05:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:23.863 19:05:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:23.863 19:05:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:23.863 19:05:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:23.863 [2024-12-05 19:05:41.263577] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:23.863 [2024-12-05 19:05:41.263716] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:23.863 [2024-12-05 19:05:41.263730] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:23.863 [2024-12-05 19:05:41.263765] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:23.863 [2024-12-05 19:05:41.267884] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000458f0 00:15:23.863 19:05:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:23.863 19:05:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:23.863 [2024-12-05 19:05:41.270038] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:24.805 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:24.805 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:24.805 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:24.805 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:24.805 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:24.805 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:24.805 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:24.805 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:24.805 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:24.805 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:24.805 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:24.805 "name": "raid_bdev1", 00:15:24.805 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:24.805 "strip_size_kb": 64, 00:15:24.805 "state": "online", 00:15:24.805 "raid_level": "raid5f", 00:15:24.805 "superblock": true, 00:15:24.805 "num_base_bdevs": 4, 00:15:24.805 "num_base_bdevs_discovered": 4, 00:15:24.805 "num_base_bdevs_operational": 4, 00:15:24.805 "process": { 00:15:24.805 "type": "rebuild", 00:15:24.805 "target": "spare", 00:15:24.805 "progress": { 00:15:24.805 "blocks": 19200, 00:15:24.805 "percent": 10 00:15:24.805 } 00:15:24.805 }, 00:15:24.805 "base_bdevs_list": [ 00:15:24.805 { 00:15:24.805 "name": "spare", 00:15:24.805 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:24.805 "is_configured": true, 00:15:24.805 "data_offset": 2048, 00:15:24.805 "data_size": 63488 00:15:24.805 }, 00:15:24.805 { 00:15:24.805 "name": "BaseBdev2", 00:15:24.805 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:24.805 "is_configured": true, 00:15:24.805 "data_offset": 2048, 00:15:24.805 "data_size": 63488 00:15:24.805 }, 00:15:24.805 { 00:15:24.805 "name": "BaseBdev3", 00:15:24.805 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:24.805 "is_configured": true, 00:15:24.805 "data_offset": 2048, 00:15:24.805 "data_size": 63488 00:15:24.805 }, 00:15:24.805 { 00:15:24.805 "name": "BaseBdev4", 00:15:24.805 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:24.805 "is_configured": true, 00:15:24.805 "data_offset": 2048, 00:15:24.805 "data_size": 63488 00:15:24.805 } 00:15:24.805 ] 00:15:24.805 }' 00:15:24.805 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:25.063 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:25.063 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:25.063 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:25.063 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:25.063 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:25.063 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:25.063 [2024-12-05 19:05:42.430552] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:25.064 [2024-12-05 19:05:42.475131] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:25.064 [2024-12-05 19:05:42.475180] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:25.064 [2024-12-05 19:05:42.475198] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:25.064 [2024-12-05 19:05:42.475205] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:25.064 "name": "raid_bdev1", 00:15:25.064 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:25.064 "strip_size_kb": 64, 00:15:25.064 "state": "online", 00:15:25.064 "raid_level": "raid5f", 00:15:25.064 "superblock": true, 00:15:25.064 "num_base_bdevs": 4, 00:15:25.064 "num_base_bdevs_discovered": 3, 00:15:25.064 "num_base_bdevs_operational": 3, 00:15:25.064 "base_bdevs_list": [ 00:15:25.064 { 00:15:25.064 "name": null, 00:15:25.064 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:25.064 "is_configured": false, 00:15:25.064 "data_offset": 0, 00:15:25.064 "data_size": 63488 00:15:25.064 }, 00:15:25.064 { 00:15:25.064 "name": "BaseBdev2", 00:15:25.064 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:25.064 "is_configured": true, 00:15:25.064 "data_offset": 2048, 00:15:25.064 "data_size": 63488 00:15:25.064 }, 00:15:25.064 { 00:15:25.064 "name": "BaseBdev3", 00:15:25.064 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:25.064 "is_configured": true, 00:15:25.064 "data_offset": 2048, 00:15:25.064 "data_size": 63488 00:15:25.064 }, 00:15:25.064 { 00:15:25.064 "name": "BaseBdev4", 00:15:25.064 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:25.064 "is_configured": true, 00:15:25.064 "data_offset": 2048, 00:15:25.064 "data_size": 63488 00:15:25.064 } 00:15:25.064 ] 00:15:25.064 }' 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:25.064 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:25.631 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:25.632 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:25.632 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:25.632 [2024-12-05 19:05:42.923437] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:25.632 [2024-12-05 19:05:42.923531] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:25.632 [2024-12-05 19:05:42.923579] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:15:25.632 [2024-12-05 19:05:42.923611] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:25.632 [2024-12-05 19:05:42.924080] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:25.632 [2024-12-05 19:05:42.924139] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:25.632 [2024-12-05 19:05:42.924245] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:25.632 [2024-12-05 19:05:42.924284] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:25.632 [2024-12-05 19:05:42.924327] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:25.632 [2024-12-05 19:05:42.924374] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:25.632 [2024-12-05 19:05:42.927614] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000459c0 00:15:25.632 spare 00:15:25.632 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:25.632 19:05:42 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:25.632 [2024-12-05 19:05:42.929822] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:26.569 19:05:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:26.569 19:05:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:26.569 19:05:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:26.569 19:05:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:26.569 19:05:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:26.569 19:05:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:26.569 19:05:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:26.569 19:05:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:26.569 19:05:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:26.569 19:05:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:26.569 19:05:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:26.569 "name": "raid_bdev1", 00:15:26.569 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:26.569 "strip_size_kb": 64, 00:15:26.569 "state": "online", 00:15:26.569 "raid_level": "raid5f", 00:15:26.569 "superblock": true, 00:15:26.569 "num_base_bdevs": 4, 00:15:26.569 "num_base_bdevs_discovered": 4, 00:15:26.569 "num_base_bdevs_operational": 4, 00:15:26.569 "process": { 00:15:26.569 "type": "rebuild", 00:15:26.569 "target": "spare", 00:15:26.569 "progress": { 00:15:26.569 "blocks": 19200, 00:15:26.569 "percent": 10 00:15:26.569 } 00:15:26.569 }, 00:15:26.569 "base_bdevs_list": [ 00:15:26.569 { 00:15:26.569 "name": "spare", 00:15:26.569 "uuid": "5d2ce5c8-f07c-5999-a04b-5da6d9906917", 00:15:26.569 "is_configured": true, 00:15:26.569 "data_offset": 2048, 00:15:26.569 "data_size": 63488 00:15:26.569 }, 00:15:26.569 { 00:15:26.569 "name": "BaseBdev2", 00:15:26.569 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:26.569 "is_configured": true, 00:15:26.569 "data_offset": 2048, 00:15:26.569 "data_size": 63488 00:15:26.569 }, 00:15:26.569 { 00:15:26.569 "name": "BaseBdev3", 00:15:26.569 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:26.569 "is_configured": true, 00:15:26.569 "data_offset": 2048, 00:15:26.569 "data_size": 63488 00:15:26.569 }, 00:15:26.569 { 00:15:26.569 "name": "BaseBdev4", 00:15:26.569 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:26.569 "is_configured": true, 00:15:26.569 "data_offset": 2048, 00:15:26.569 "data_size": 63488 00:15:26.569 } 00:15:26.569 ] 00:15:26.569 }' 00:15:26.569 19:05:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:26.569 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:26.569 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:26.570 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:26.570 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:26.570 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:26.570 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:26.570 [2024-12-05 19:05:44.078261] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:26.829 [2024-12-05 19:05:44.134794] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:26.829 [2024-12-05 19:05:44.134892] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:26.829 [2024-12-05 19:05:44.134928] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:26.829 [2024-12-05 19:05:44.134950] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:26.829 "name": "raid_bdev1", 00:15:26.829 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:26.829 "strip_size_kb": 64, 00:15:26.829 "state": "online", 00:15:26.829 "raid_level": "raid5f", 00:15:26.829 "superblock": true, 00:15:26.829 "num_base_bdevs": 4, 00:15:26.829 "num_base_bdevs_discovered": 3, 00:15:26.829 "num_base_bdevs_operational": 3, 00:15:26.829 "base_bdevs_list": [ 00:15:26.829 { 00:15:26.829 "name": null, 00:15:26.829 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:26.829 "is_configured": false, 00:15:26.829 "data_offset": 0, 00:15:26.829 "data_size": 63488 00:15:26.829 }, 00:15:26.829 { 00:15:26.829 "name": "BaseBdev2", 00:15:26.829 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:26.829 "is_configured": true, 00:15:26.829 "data_offset": 2048, 00:15:26.829 "data_size": 63488 00:15:26.829 }, 00:15:26.829 { 00:15:26.829 "name": "BaseBdev3", 00:15:26.829 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:26.829 "is_configured": true, 00:15:26.829 "data_offset": 2048, 00:15:26.829 "data_size": 63488 00:15:26.829 }, 00:15:26.829 { 00:15:26.829 "name": "BaseBdev4", 00:15:26.829 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:26.829 "is_configured": true, 00:15:26.829 "data_offset": 2048, 00:15:26.829 "data_size": 63488 00:15:26.829 } 00:15:26.829 ] 00:15:26.829 }' 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:26.829 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:27.088 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:27.088 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:27.089 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:27.089 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:27.089 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:27.089 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.089 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.089 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:27.089 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:27.089 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.089 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:27.089 "name": "raid_bdev1", 00:15:27.089 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:27.089 "strip_size_kb": 64, 00:15:27.089 "state": "online", 00:15:27.089 "raid_level": "raid5f", 00:15:27.089 "superblock": true, 00:15:27.089 "num_base_bdevs": 4, 00:15:27.089 "num_base_bdevs_discovered": 3, 00:15:27.089 "num_base_bdevs_operational": 3, 00:15:27.089 "base_bdevs_list": [ 00:15:27.089 { 00:15:27.089 "name": null, 00:15:27.089 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:27.089 "is_configured": false, 00:15:27.089 "data_offset": 0, 00:15:27.089 "data_size": 63488 00:15:27.089 }, 00:15:27.089 { 00:15:27.089 "name": "BaseBdev2", 00:15:27.089 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:27.089 "is_configured": true, 00:15:27.089 "data_offset": 2048, 00:15:27.089 "data_size": 63488 00:15:27.089 }, 00:15:27.089 { 00:15:27.089 "name": "BaseBdev3", 00:15:27.089 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:27.089 "is_configured": true, 00:15:27.089 "data_offset": 2048, 00:15:27.089 "data_size": 63488 00:15:27.089 }, 00:15:27.089 { 00:15:27.089 "name": "BaseBdev4", 00:15:27.089 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:27.089 "is_configured": true, 00:15:27.089 "data_offset": 2048, 00:15:27.089 "data_size": 63488 00:15:27.089 } 00:15:27.089 ] 00:15:27.089 }' 00:15:27.089 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:27.348 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:27.348 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:27.348 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:27.348 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:27.348 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.348 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:27.348 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.348 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:27.348 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:27.348 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:27.348 [2024-12-05 19:05:44.742841] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:27.348 [2024-12-05 19:05:44.742891] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:27.348 [2024-12-05 19:05:44.742910] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:15:27.348 [2024-12-05 19:05:44.742920] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:27.348 [2024-12-05 19:05:44.743300] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:27.348 [2024-12-05 19:05:44.743320] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:27.348 [2024-12-05 19:05:44.743382] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:27.348 [2024-12-05 19:05:44.743400] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:27.348 [2024-12-05 19:05:44.743407] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:27.348 [2024-12-05 19:05:44.743418] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:27.348 BaseBdev1 00:15:27.348 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:27.348 19:05:44 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:28.286 "name": "raid_bdev1", 00:15:28.286 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:28.286 "strip_size_kb": 64, 00:15:28.286 "state": "online", 00:15:28.286 "raid_level": "raid5f", 00:15:28.286 "superblock": true, 00:15:28.286 "num_base_bdevs": 4, 00:15:28.286 "num_base_bdevs_discovered": 3, 00:15:28.286 "num_base_bdevs_operational": 3, 00:15:28.286 "base_bdevs_list": [ 00:15:28.286 { 00:15:28.286 "name": null, 00:15:28.286 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:28.286 "is_configured": false, 00:15:28.286 "data_offset": 0, 00:15:28.286 "data_size": 63488 00:15:28.286 }, 00:15:28.286 { 00:15:28.286 "name": "BaseBdev2", 00:15:28.286 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:28.286 "is_configured": true, 00:15:28.286 "data_offset": 2048, 00:15:28.286 "data_size": 63488 00:15:28.286 }, 00:15:28.286 { 00:15:28.286 "name": "BaseBdev3", 00:15:28.286 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:28.286 "is_configured": true, 00:15:28.286 "data_offset": 2048, 00:15:28.286 "data_size": 63488 00:15:28.286 }, 00:15:28.286 { 00:15:28.286 "name": "BaseBdev4", 00:15:28.286 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:28.286 "is_configured": true, 00:15:28.286 "data_offset": 2048, 00:15:28.286 "data_size": 63488 00:15:28.286 } 00:15:28.286 ] 00:15:28.286 }' 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:28.286 19:05:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:28.865 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:28.866 "name": "raid_bdev1", 00:15:28.866 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:28.866 "strip_size_kb": 64, 00:15:28.866 "state": "online", 00:15:28.866 "raid_level": "raid5f", 00:15:28.866 "superblock": true, 00:15:28.866 "num_base_bdevs": 4, 00:15:28.866 "num_base_bdevs_discovered": 3, 00:15:28.866 "num_base_bdevs_operational": 3, 00:15:28.866 "base_bdevs_list": [ 00:15:28.866 { 00:15:28.866 "name": null, 00:15:28.866 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:28.866 "is_configured": false, 00:15:28.866 "data_offset": 0, 00:15:28.866 "data_size": 63488 00:15:28.866 }, 00:15:28.866 { 00:15:28.866 "name": "BaseBdev2", 00:15:28.866 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:28.866 "is_configured": true, 00:15:28.866 "data_offset": 2048, 00:15:28.866 "data_size": 63488 00:15:28.866 }, 00:15:28.866 { 00:15:28.866 "name": "BaseBdev3", 00:15:28.866 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:28.866 "is_configured": true, 00:15:28.866 "data_offset": 2048, 00:15:28.866 "data_size": 63488 00:15:28.866 }, 00:15:28.866 { 00:15:28.866 "name": "BaseBdev4", 00:15:28.866 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:28.866 "is_configured": true, 00:15:28.866 "data_offset": 2048, 00:15:28.866 "data_size": 63488 00:15:28.866 } 00:15:28.866 ] 00:15:28.866 }' 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # local es=0 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:28.866 [2024-12-05 19:05:46.368097] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:28.866 [2024-12-05 19:05:46.368211] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:28.866 [2024-12-05 19:05:46.368222] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:28.866 request: 00:15:28.866 { 00:15:28.866 "base_bdev": "BaseBdev1", 00:15:28.866 "raid_bdev": "raid_bdev1", 00:15:28.866 "method": "bdev_raid_add_base_bdev", 00:15:28.866 "req_id": 1 00:15:28.866 } 00:15:28.866 Got JSON-RPC error response 00:15:28.866 response: 00:15:28.866 { 00:15:28.866 "code": -22, 00:15:28.866 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:28.866 } 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@655 -- # es=1 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:28.866 19:05:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:30.249 "name": "raid_bdev1", 00:15:30.249 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:30.249 "strip_size_kb": 64, 00:15:30.249 "state": "online", 00:15:30.249 "raid_level": "raid5f", 00:15:30.249 "superblock": true, 00:15:30.249 "num_base_bdevs": 4, 00:15:30.249 "num_base_bdevs_discovered": 3, 00:15:30.249 "num_base_bdevs_operational": 3, 00:15:30.249 "base_bdevs_list": [ 00:15:30.249 { 00:15:30.249 "name": null, 00:15:30.249 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:30.249 "is_configured": false, 00:15:30.249 "data_offset": 0, 00:15:30.249 "data_size": 63488 00:15:30.249 }, 00:15:30.249 { 00:15:30.249 "name": "BaseBdev2", 00:15:30.249 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:30.249 "is_configured": true, 00:15:30.249 "data_offset": 2048, 00:15:30.249 "data_size": 63488 00:15:30.249 }, 00:15:30.249 { 00:15:30.249 "name": "BaseBdev3", 00:15:30.249 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:30.249 "is_configured": true, 00:15:30.249 "data_offset": 2048, 00:15:30.249 "data_size": 63488 00:15:30.249 }, 00:15:30.249 { 00:15:30.249 "name": "BaseBdev4", 00:15:30.249 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:30.249 "is_configured": true, 00:15:30.249 "data_offset": 2048, 00:15:30.249 "data_size": 63488 00:15:30.249 } 00:15:30.249 ] 00:15:30.249 }' 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:30.249 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:30.510 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:30.510 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:30.510 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:30.510 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:30.510 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:30.510 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:30.510 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:30.510 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:30.510 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:30.510 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:30.510 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:30.510 "name": "raid_bdev1", 00:15:30.510 "uuid": "0d666353-9de4-4f70-be8b-16da06e72d28", 00:15:30.510 "strip_size_kb": 64, 00:15:30.510 "state": "online", 00:15:30.510 "raid_level": "raid5f", 00:15:30.510 "superblock": true, 00:15:30.510 "num_base_bdevs": 4, 00:15:30.510 "num_base_bdevs_discovered": 3, 00:15:30.510 "num_base_bdevs_operational": 3, 00:15:30.510 "base_bdevs_list": [ 00:15:30.510 { 00:15:30.510 "name": null, 00:15:30.510 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:30.510 "is_configured": false, 00:15:30.510 "data_offset": 0, 00:15:30.510 "data_size": 63488 00:15:30.510 }, 00:15:30.510 { 00:15:30.510 "name": "BaseBdev2", 00:15:30.510 "uuid": "2a542775-a668-5744-9498-b7d6271e7af3", 00:15:30.510 "is_configured": true, 00:15:30.510 "data_offset": 2048, 00:15:30.510 "data_size": 63488 00:15:30.510 }, 00:15:30.510 { 00:15:30.510 "name": "BaseBdev3", 00:15:30.510 "uuid": "8066f925-26b2-57f2-a235-5b87194bd2af", 00:15:30.510 "is_configured": true, 00:15:30.510 "data_offset": 2048, 00:15:30.510 "data_size": 63488 00:15:30.510 }, 00:15:30.510 { 00:15:30.510 "name": "BaseBdev4", 00:15:30.510 "uuid": "15c90b7b-ac46-59bb-a140-6e94c0b63ecc", 00:15:30.510 "is_configured": true, 00:15:30.510 "data_offset": 2048, 00:15:30.510 "data_size": 63488 00:15:30.510 } 00:15:30.510 ] 00:15:30.510 }' 00:15:30.510 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:30.510 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:30.510 19:05:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:30.510 19:05:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:30.510 19:05:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 92849 00:15:30.510 19:05:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # '[' -z 92849 ']' 00:15:30.510 19:05:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@958 -- # kill -0 92849 00:15:30.510 19:05:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # uname 00:15:30.510 19:05:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:30.510 19:05:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 92849 00:15:30.510 19:05:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:30.510 19:05:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:30.510 19:05:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@972 -- # echo 'killing process with pid 92849' 00:15:30.510 killing process with pid 92849 00:15:30.510 19:05:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@973 -- # kill 92849 00:15:30.510 Received shutdown signal, test time was about 60.000000 seconds 00:15:30.510 00:15:30.510 Latency(us) 00:15:30.510 [2024-12-05T19:05:48.069Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:30.510 [2024-12-05T19:05:48.069Z] =================================================================================================================== 00:15:30.510 [2024-12-05T19:05:48.069Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:30.510 [2024-12-05 19:05:48.061411] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:30.510 [2024-12-05 19:05:48.061552] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:30.510 19:05:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@978 -- # wait 92849 00:15:30.510 [2024-12-05 19:05:48.061656] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:30.510 [2024-12-05 19:05:48.061712] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:15:30.770 [2024-12-05 19:05:48.111487] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:30.770 19:05:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:15:30.770 00:15:30.770 real 0m25.348s 00:15:30.770 user 0m32.180s 00:15:30.770 sys 0m3.265s 00:15:30.770 19:05:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:30.771 19:05:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:30.771 ************************************ 00:15:30.771 END TEST raid5f_rebuild_test_sb 00:15:30.771 ************************************ 00:15:31.031 19:05:48 bdev_raid -- bdev/bdev_raid.sh@995 -- # base_blocklen=4096 00:15:31.031 19:05:48 bdev_raid -- bdev/bdev_raid.sh@997 -- # run_test raid_state_function_test_sb_4k raid_state_function_test raid1 2 true 00:15:31.031 19:05:48 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:15:31.031 19:05:48 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:31.031 19:05:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:31.031 ************************************ 00:15:31.031 START TEST raid_state_function_test_sb_4k 00:15:31.031 ************************************ 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@229 -- # raid_pid=93497 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 93497' 00:15:31.031 Process raid pid: 93497 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@231 -- # waitforlisten 93497 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@835 -- # '[' -z 93497 ']' 00:15:31.031 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:31.031 19:05:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:31.031 [2024-12-05 19:05:48.485992] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:15:31.031 [2024-12-05 19:05:48.486125] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:31.292 [2024-12-05 19:05:48.643585] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:31.292 [2024-12-05 19:05:48.669410] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:31.292 [2024-12-05 19:05:48.712882] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:31.292 [2024-12-05 19:05:48.712923] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@868 -- # return 0 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:31.862 [2024-12-05 19:05:49.300128] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:31.862 [2024-12-05 19:05:49.300185] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:31.862 [2024-12-05 19:05:49.300194] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:31.862 [2024-12-05 19:05:49.300203] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:31.862 "name": "Existed_Raid", 00:15:31.862 "uuid": "0c3096e9-761c-41bf-a257-2679711d2706", 00:15:31.862 "strip_size_kb": 0, 00:15:31.862 "state": "configuring", 00:15:31.862 "raid_level": "raid1", 00:15:31.862 "superblock": true, 00:15:31.862 "num_base_bdevs": 2, 00:15:31.862 "num_base_bdevs_discovered": 0, 00:15:31.862 "num_base_bdevs_operational": 2, 00:15:31.862 "base_bdevs_list": [ 00:15:31.862 { 00:15:31.862 "name": "BaseBdev1", 00:15:31.862 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:31.862 "is_configured": false, 00:15:31.862 "data_offset": 0, 00:15:31.862 "data_size": 0 00:15:31.862 }, 00:15:31.862 { 00:15:31.862 "name": "BaseBdev2", 00:15:31.862 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:31.862 "is_configured": false, 00:15:31.862 "data_offset": 0, 00:15:31.862 "data_size": 0 00:15:31.862 } 00:15:31.862 ] 00:15:31.862 }' 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:31.862 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:32.433 [2024-12-05 19:05:49.747338] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:32.433 [2024-12-05 19:05:49.747418] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:32.433 [2024-12-05 19:05:49.759313] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:32.433 [2024-12-05 19:05:49.759400] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:32.433 [2024-12-05 19:05:49.759425] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:32.433 [2024-12-05 19:05:49.759459] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:32.433 [2024-12-05 19:05:49.780343] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:32.433 BaseBdev1 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # local i 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:32.433 [ 00:15:32.433 { 00:15:32.433 "name": "BaseBdev1", 00:15:32.433 "aliases": [ 00:15:32.433 "bfd55c1d-2082-4abb-8504-155735e9940b" 00:15:32.433 ], 00:15:32.433 "product_name": "Malloc disk", 00:15:32.433 "block_size": 4096, 00:15:32.433 "num_blocks": 8192, 00:15:32.433 "uuid": "bfd55c1d-2082-4abb-8504-155735e9940b", 00:15:32.433 "assigned_rate_limits": { 00:15:32.433 "rw_ios_per_sec": 0, 00:15:32.433 "rw_mbytes_per_sec": 0, 00:15:32.433 "r_mbytes_per_sec": 0, 00:15:32.433 "w_mbytes_per_sec": 0 00:15:32.433 }, 00:15:32.433 "claimed": true, 00:15:32.433 "claim_type": "exclusive_write", 00:15:32.433 "zoned": false, 00:15:32.433 "supported_io_types": { 00:15:32.433 "read": true, 00:15:32.433 "write": true, 00:15:32.433 "unmap": true, 00:15:32.433 "flush": true, 00:15:32.433 "reset": true, 00:15:32.433 "nvme_admin": false, 00:15:32.433 "nvme_io": false, 00:15:32.433 "nvme_io_md": false, 00:15:32.433 "write_zeroes": true, 00:15:32.433 "zcopy": true, 00:15:32.433 "get_zone_info": false, 00:15:32.433 "zone_management": false, 00:15:32.433 "zone_append": false, 00:15:32.433 "compare": false, 00:15:32.433 "compare_and_write": false, 00:15:32.433 "abort": true, 00:15:32.433 "seek_hole": false, 00:15:32.433 "seek_data": false, 00:15:32.433 "copy": true, 00:15:32.433 "nvme_iov_md": false 00:15:32.433 }, 00:15:32.433 "memory_domains": [ 00:15:32.433 { 00:15:32.433 "dma_device_id": "system", 00:15:32.433 "dma_device_type": 1 00:15:32.433 }, 00:15:32.433 { 00:15:32.433 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:32.433 "dma_device_type": 2 00:15:32.433 } 00:15:32.433 ], 00:15:32.433 "driver_specific": {} 00:15:32.433 } 00:15:32.433 ] 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@911 -- # return 0 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:32.433 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:32.434 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:32.434 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:32.434 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:32.434 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:32.434 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:32.434 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:32.434 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:32.434 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.434 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:32.434 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:32.434 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:32.434 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:32.434 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:32.434 "name": "Existed_Raid", 00:15:32.434 "uuid": "b06054d0-d5d9-4685-bd3e-2e8c4cca2c0c", 00:15:32.434 "strip_size_kb": 0, 00:15:32.434 "state": "configuring", 00:15:32.434 "raid_level": "raid1", 00:15:32.434 "superblock": true, 00:15:32.434 "num_base_bdevs": 2, 00:15:32.434 "num_base_bdevs_discovered": 1, 00:15:32.434 "num_base_bdevs_operational": 2, 00:15:32.434 "base_bdevs_list": [ 00:15:32.434 { 00:15:32.434 "name": "BaseBdev1", 00:15:32.434 "uuid": "bfd55c1d-2082-4abb-8504-155735e9940b", 00:15:32.434 "is_configured": true, 00:15:32.434 "data_offset": 256, 00:15:32.434 "data_size": 7936 00:15:32.434 }, 00:15:32.434 { 00:15:32.434 "name": "BaseBdev2", 00:15:32.434 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:32.434 "is_configured": false, 00:15:32.434 "data_offset": 0, 00:15:32.434 "data_size": 0 00:15:32.434 } 00:15:32.434 ] 00:15:32.434 }' 00:15:32.434 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:32.434 19:05:49 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.004 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:33.004 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.004 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.004 [2024-12-05 19:05:50.295497] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:33.004 [2024-12-05 19:05:50.295542] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:15:33.004 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.004 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:33.004 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.004 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.005 [2024-12-05 19:05:50.307498] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:33.005 [2024-12-05 19:05:50.309396] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:33.005 [2024-12-05 19:05:50.309451] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:33.005 "name": "Existed_Raid", 00:15:33.005 "uuid": "d8e3c20b-1704-4411-bb10-3650d433c439", 00:15:33.005 "strip_size_kb": 0, 00:15:33.005 "state": "configuring", 00:15:33.005 "raid_level": "raid1", 00:15:33.005 "superblock": true, 00:15:33.005 "num_base_bdevs": 2, 00:15:33.005 "num_base_bdevs_discovered": 1, 00:15:33.005 "num_base_bdevs_operational": 2, 00:15:33.005 "base_bdevs_list": [ 00:15:33.005 { 00:15:33.005 "name": "BaseBdev1", 00:15:33.005 "uuid": "bfd55c1d-2082-4abb-8504-155735e9940b", 00:15:33.005 "is_configured": true, 00:15:33.005 "data_offset": 256, 00:15:33.005 "data_size": 7936 00:15:33.005 }, 00:15:33.005 { 00:15:33.005 "name": "BaseBdev2", 00:15:33.005 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:33.005 "is_configured": false, 00:15:33.005 "data_offset": 0, 00:15:33.005 "data_size": 0 00:15:33.005 } 00:15:33.005 ] 00:15:33.005 }' 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:33.005 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.265 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2 00:15:33.265 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.265 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.266 [2024-12-05 19:05:50.733872] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:33.266 [2024-12-05 19:05:50.734077] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:33.266 [2024-12-05 19:05:50.734095] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:33.266 BaseBdev2 00:15:33.266 [2024-12-05 19:05:50.734377] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:15:33.266 [2024-12-05 19:05:50.734531] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:33.266 [2024-12-05 19:05:50.734554] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:15:33.266 [2024-12-05 19:05:50.734702] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@905 -- # local i 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.266 [ 00:15:33.266 { 00:15:33.266 "name": "BaseBdev2", 00:15:33.266 "aliases": [ 00:15:33.266 "df18dc1f-f596-49cd-8582-a98a3bff7870" 00:15:33.266 ], 00:15:33.266 "product_name": "Malloc disk", 00:15:33.266 "block_size": 4096, 00:15:33.266 "num_blocks": 8192, 00:15:33.266 "uuid": "df18dc1f-f596-49cd-8582-a98a3bff7870", 00:15:33.266 "assigned_rate_limits": { 00:15:33.266 "rw_ios_per_sec": 0, 00:15:33.266 "rw_mbytes_per_sec": 0, 00:15:33.266 "r_mbytes_per_sec": 0, 00:15:33.266 "w_mbytes_per_sec": 0 00:15:33.266 }, 00:15:33.266 "claimed": true, 00:15:33.266 "claim_type": "exclusive_write", 00:15:33.266 "zoned": false, 00:15:33.266 "supported_io_types": { 00:15:33.266 "read": true, 00:15:33.266 "write": true, 00:15:33.266 "unmap": true, 00:15:33.266 "flush": true, 00:15:33.266 "reset": true, 00:15:33.266 "nvme_admin": false, 00:15:33.266 "nvme_io": false, 00:15:33.266 "nvme_io_md": false, 00:15:33.266 "write_zeroes": true, 00:15:33.266 "zcopy": true, 00:15:33.266 "get_zone_info": false, 00:15:33.266 "zone_management": false, 00:15:33.266 "zone_append": false, 00:15:33.266 "compare": false, 00:15:33.266 "compare_and_write": false, 00:15:33.266 "abort": true, 00:15:33.266 "seek_hole": false, 00:15:33.266 "seek_data": false, 00:15:33.266 "copy": true, 00:15:33.266 "nvme_iov_md": false 00:15:33.266 }, 00:15:33.266 "memory_domains": [ 00:15:33.266 { 00:15:33.266 "dma_device_id": "system", 00:15:33.266 "dma_device_type": 1 00:15:33.266 }, 00:15:33.266 { 00:15:33.266 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:33.266 "dma_device_type": 2 00:15:33.266 } 00:15:33.266 ], 00:15:33.266 "driver_specific": {} 00:15:33.266 } 00:15:33.266 ] 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@911 -- # return 0 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:33.266 "name": "Existed_Raid", 00:15:33.266 "uuid": "d8e3c20b-1704-4411-bb10-3650d433c439", 00:15:33.266 "strip_size_kb": 0, 00:15:33.266 "state": "online", 00:15:33.266 "raid_level": "raid1", 00:15:33.266 "superblock": true, 00:15:33.266 "num_base_bdevs": 2, 00:15:33.266 "num_base_bdevs_discovered": 2, 00:15:33.266 "num_base_bdevs_operational": 2, 00:15:33.266 "base_bdevs_list": [ 00:15:33.266 { 00:15:33.266 "name": "BaseBdev1", 00:15:33.266 "uuid": "bfd55c1d-2082-4abb-8504-155735e9940b", 00:15:33.266 "is_configured": true, 00:15:33.266 "data_offset": 256, 00:15:33.266 "data_size": 7936 00:15:33.266 }, 00:15:33.266 { 00:15:33.266 "name": "BaseBdev2", 00:15:33.266 "uuid": "df18dc1f-f596-49cd-8582-a98a3bff7870", 00:15:33.266 "is_configured": true, 00:15:33.266 "data_offset": 256, 00:15:33.266 "data_size": 7936 00:15:33.266 } 00:15:33.266 ] 00:15:33.266 }' 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:33.266 19:05:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:33.836 [2024-12-05 19:05:51.197353] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:33.836 "name": "Existed_Raid", 00:15:33.836 "aliases": [ 00:15:33.836 "d8e3c20b-1704-4411-bb10-3650d433c439" 00:15:33.836 ], 00:15:33.836 "product_name": "Raid Volume", 00:15:33.836 "block_size": 4096, 00:15:33.836 "num_blocks": 7936, 00:15:33.836 "uuid": "d8e3c20b-1704-4411-bb10-3650d433c439", 00:15:33.836 "assigned_rate_limits": { 00:15:33.836 "rw_ios_per_sec": 0, 00:15:33.836 "rw_mbytes_per_sec": 0, 00:15:33.836 "r_mbytes_per_sec": 0, 00:15:33.836 "w_mbytes_per_sec": 0 00:15:33.836 }, 00:15:33.836 "claimed": false, 00:15:33.836 "zoned": false, 00:15:33.836 "supported_io_types": { 00:15:33.836 "read": true, 00:15:33.836 "write": true, 00:15:33.836 "unmap": false, 00:15:33.836 "flush": false, 00:15:33.836 "reset": true, 00:15:33.836 "nvme_admin": false, 00:15:33.836 "nvme_io": false, 00:15:33.836 "nvme_io_md": false, 00:15:33.836 "write_zeroes": true, 00:15:33.836 "zcopy": false, 00:15:33.836 "get_zone_info": false, 00:15:33.836 "zone_management": false, 00:15:33.836 "zone_append": false, 00:15:33.836 "compare": false, 00:15:33.836 "compare_and_write": false, 00:15:33.836 "abort": false, 00:15:33.836 "seek_hole": false, 00:15:33.836 "seek_data": false, 00:15:33.836 "copy": false, 00:15:33.836 "nvme_iov_md": false 00:15:33.836 }, 00:15:33.836 "memory_domains": [ 00:15:33.836 { 00:15:33.836 "dma_device_id": "system", 00:15:33.836 "dma_device_type": 1 00:15:33.836 }, 00:15:33.836 { 00:15:33.836 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:33.836 "dma_device_type": 2 00:15:33.836 }, 00:15:33.836 { 00:15:33.836 "dma_device_id": "system", 00:15:33.836 "dma_device_type": 1 00:15:33.836 }, 00:15:33.836 { 00:15:33.836 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:33.836 "dma_device_type": 2 00:15:33.836 } 00:15:33.836 ], 00:15:33.836 "driver_specific": { 00:15:33.836 "raid": { 00:15:33.836 "uuid": "d8e3c20b-1704-4411-bb10-3650d433c439", 00:15:33.836 "strip_size_kb": 0, 00:15:33.836 "state": "online", 00:15:33.836 "raid_level": "raid1", 00:15:33.836 "superblock": true, 00:15:33.836 "num_base_bdevs": 2, 00:15:33.836 "num_base_bdevs_discovered": 2, 00:15:33.836 "num_base_bdevs_operational": 2, 00:15:33.836 "base_bdevs_list": [ 00:15:33.836 { 00:15:33.836 "name": "BaseBdev1", 00:15:33.836 "uuid": "bfd55c1d-2082-4abb-8504-155735e9940b", 00:15:33.836 "is_configured": true, 00:15:33.836 "data_offset": 256, 00:15:33.836 "data_size": 7936 00:15:33.836 }, 00:15:33.836 { 00:15:33.836 "name": "BaseBdev2", 00:15:33.836 "uuid": "df18dc1f-f596-49cd-8582-a98a3bff7870", 00:15:33.836 "is_configured": true, 00:15:33.836 "data_offset": 256, 00:15:33.836 "data_size": 7936 00:15:33.836 } 00:15:33.836 ] 00:15:33.836 } 00:15:33.836 } 00:15:33.836 }' 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:33.836 BaseBdev2' 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:33.836 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:33.837 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:33.837 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:33.837 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:33.837 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:33.837 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:33.837 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.097 [2024-12-05 19:05:51.404809] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:34.097 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:34.098 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:34.098 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:34.098 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.098 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.098 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:34.098 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.098 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.098 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:34.098 "name": "Existed_Raid", 00:15:34.098 "uuid": "d8e3c20b-1704-4411-bb10-3650d433c439", 00:15:34.098 "strip_size_kb": 0, 00:15:34.098 "state": "online", 00:15:34.098 "raid_level": "raid1", 00:15:34.098 "superblock": true, 00:15:34.098 "num_base_bdevs": 2, 00:15:34.098 "num_base_bdevs_discovered": 1, 00:15:34.098 "num_base_bdevs_operational": 1, 00:15:34.098 "base_bdevs_list": [ 00:15:34.098 { 00:15:34.098 "name": null, 00:15:34.098 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:34.098 "is_configured": false, 00:15:34.098 "data_offset": 0, 00:15:34.098 "data_size": 7936 00:15:34.098 }, 00:15:34.098 { 00:15:34.098 "name": "BaseBdev2", 00:15:34.098 "uuid": "df18dc1f-f596-49cd-8582-a98a3bff7870", 00:15:34.098 "is_configured": true, 00:15:34.098 "data_offset": 256, 00:15:34.098 "data_size": 7936 00:15:34.098 } 00:15:34.098 ] 00:15:34.098 }' 00:15:34.098 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:34.098 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.358 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:34.358 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:34.358 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.358 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.358 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.358 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:34.358 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.358 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:34.358 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:34.358 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:34.358 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.358 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.358 [2024-12-05 19:05:51.915144] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:34.358 [2024-12-05 19:05:51.915234] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:34.619 [2024-12-05 19:05:51.926513] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:34.619 [2024-12-05 19:05:51.926597] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:34.619 [2024-12-05 19:05:51.926608] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@326 -- # killprocess 93497 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@954 -- # '[' -z 93497 ']' 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@958 -- # kill -0 93497 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@959 -- # uname 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:34.619 19:05:51 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 93497 00:15:34.619 19:05:52 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:34.619 19:05:52 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:34.619 killing process with pid 93497 00:15:34.619 19:05:52 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 93497' 00:15:34.619 19:05:52 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@973 -- # kill 93497 00:15:34.619 [2024-12-05 19:05:52.012250] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:34.619 19:05:52 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@978 -- # wait 93497 00:15:34.619 [2024-12-05 19:05:52.013213] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:34.880 19:05:52 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@328 -- # return 0 00:15:34.880 00:15:34.880 real 0m3.831s 00:15:34.880 user 0m6.047s 00:15:34.880 sys 0m0.833s 00:15:34.880 19:05:52 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:34.880 19:05:52 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.880 ************************************ 00:15:34.880 END TEST raid_state_function_test_sb_4k 00:15:34.880 ************************************ 00:15:34.880 19:05:52 bdev_raid -- bdev/bdev_raid.sh@998 -- # run_test raid_superblock_test_4k raid_superblock_test raid1 2 00:15:34.880 19:05:52 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:15:34.880 19:05:52 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:34.880 19:05:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:34.880 ************************************ 00:15:34.880 START TEST raid_superblock_test_4k 00:15:34.880 ************************************ 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@399 -- # local strip_size 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@412 -- # raid_pid=93714 00:15:34.880 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:15:34.881 19:05:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@413 -- # waitforlisten 93714 00:15:34.881 19:05:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@835 -- # '[' -z 93714 ']' 00:15:34.881 19:05:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:34.881 19:05:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:34.881 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:34.881 19:05:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:34.881 19:05:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:34.881 19:05:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.881 [2024-12-05 19:05:52.392300] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:15:34.881 [2024-12-05 19:05:52.392925] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid93714 ] 00:15:35.141 [2024-12-05 19:05:52.547697] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:35.141 [2024-12-05 19:05:52.573461] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:35.141 [2024-12-05 19:05:52.616864] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:35.141 [2024-12-05 19:05:52.616903] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@868 -- # return 0 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc1 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:35.712 malloc1 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:35.712 [2024-12-05 19:05:53.232702] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:35.712 [2024-12-05 19:05:53.232759] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:35.712 [2024-12-05 19:05:53.232784] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:35.712 [2024-12-05 19:05:53.232798] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:35.712 [2024-12-05 19:05:53.234858] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:35.712 [2024-12-05 19:05:53.234893] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:35.712 pt1 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc2 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:35.712 malloc2 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:35.712 [2024-12-05 19:05:53.261356] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:35.712 [2024-12-05 19:05:53.261405] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:35.712 [2024-12-05 19:05:53.261422] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:35.712 [2024-12-05 19:05:53.261433] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:35.712 [2024-12-05 19:05:53.263452] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:35.712 [2024-12-05 19:05:53.263481] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:35.712 pt2 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:35.712 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:35.973 [2024-12-05 19:05:53.273363] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:35.973 [2024-12-05 19:05:53.275170] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:35.973 [2024-12-05 19:05:53.275308] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:35.973 [2024-12-05 19:05:53.275327] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:35.973 [2024-12-05 19:05:53.275577] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:15:35.973 [2024-12-05 19:05:53.275733] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:35.973 [2024-12-05 19:05:53.275751] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:35.973 [2024-12-05 19:05:53.275879] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:35.973 "name": "raid_bdev1", 00:15:35.973 "uuid": "684ac437-bf68-4c71-8f0e-006d4fccec01", 00:15:35.973 "strip_size_kb": 0, 00:15:35.973 "state": "online", 00:15:35.973 "raid_level": "raid1", 00:15:35.973 "superblock": true, 00:15:35.973 "num_base_bdevs": 2, 00:15:35.973 "num_base_bdevs_discovered": 2, 00:15:35.973 "num_base_bdevs_operational": 2, 00:15:35.973 "base_bdevs_list": [ 00:15:35.973 { 00:15:35.973 "name": "pt1", 00:15:35.973 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:35.973 "is_configured": true, 00:15:35.973 "data_offset": 256, 00:15:35.973 "data_size": 7936 00:15:35.973 }, 00:15:35.973 { 00:15:35.973 "name": "pt2", 00:15:35.973 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:35.973 "is_configured": true, 00:15:35.973 "data_offset": 256, 00:15:35.973 "data_size": 7936 00:15:35.973 } 00:15:35.973 ] 00:15:35.973 }' 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:35.973 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.231 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:15:36.231 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:36.231 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:36.231 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:36.231 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:36.231 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:36.231 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:36.231 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.231 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.231 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:36.231 [2024-12-05 19:05:53.748810] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:36.231 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:36.498 "name": "raid_bdev1", 00:15:36.498 "aliases": [ 00:15:36.498 "684ac437-bf68-4c71-8f0e-006d4fccec01" 00:15:36.498 ], 00:15:36.498 "product_name": "Raid Volume", 00:15:36.498 "block_size": 4096, 00:15:36.498 "num_blocks": 7936, 00:15:36.498 "uuid": "684ac437-bf68-4c71-8f0e-006d4fccec01", 00:15:36.498 "assigned_rate_limits": { 00:15:36.498 "rw_ios_per_sec": 0, 00:15:36.498 "rw_mbytes_per_sec": 0, 00:15:36.498 "r_mbytes_per_sec": 0, 00:15:36.498 "w_mbytes_per_sec": 0 00:15:36.498 }, 00:15:36.498 "claimed": false, 00:15:36.498 "zoned": false, 00:15:36.498 "supported_io_types": { 00:15:36.498 "read": true, 00:15:36.498 "write": true, 00:15:36.498 "unmap": false, 00:15:36.498 "flush": false, 00:15:36.498 "reset": true, 00:15:36.498 "nvme_admin": false, 00:15:36.498 "nvme_io": false, 00:15:36.498 "nvme_io_md": false, 00:15:36.498 "write_zeroes": true, 00:15:36.498 "zcopy": false, 00:15:36.498 "get_zone_info": false, 00:15:36.498 "zone_management": false, 00:15:36.498 "zone_append": false, 00:15:36.498 "compare": false, 00:15:36.498 "compare_and_write": false, 00:15:36.498 "abort": false, 00:15:36.498 "seek_hole": false, 00:15:36.498 "seek_data": false, 00:15:36.498 "copy": false, 00:15:36.498 "nvme_iov_md": false 00:15:36.498 }, 00:15:36.498 "memory_domains": [ 00:15:36.498 { 00:15:36.498 "dma_device_id": "system", 00:15:36.498 "dma_device_type": 1 00:15:36.498 }, 00:15:36.498 { 00:15:36.498 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:36.498 "dma_device_type": 2 00:15:36.498 }, 00:15:36.498 { 00:15:36.498 "dma_device_id": "system", 00:15:36.498 "dma_device_type": 1 00:15:36.498 }, 00:15:36.498 { 00:15:36.498 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:36.498 "dma_device_type": 2 00:15:36.498 } 00:15:36.498 ], 00:15:36.498 "driver_specific": { 00:15:36.498 "raid": { 00:15:36.498 "uuid": "684ac437-bf68-4c71-8f0e-006d4fccec01", 00:15:36.498 "strip_size_kb": 0, 00:15:36.498 "state": "online", 00:15:36.498 "raid_level": "raid1", 00:15:36.498 "superblock": true, 00:15:36.498 "num_base_bdevs": 2, 00:15:36.498 "num_base_bdevs_discovered": 2, 00:15:36.498 "num_base_bdevs_operational": 2, 00:15:36.498 "base_bdevs_list": [ 00:15:36.498 { 00:15:36.498 "name": "pt1", 00:15:36.498 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:36.498 "is_configured": true, 00:15:36.498 "data_offset": 256, 00:15:36.498 "data_size": 7936 00:15:36.498 }, 00:15:36.498 { 00:15:36.498 "name": "pt2", 00:15:36.498 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:36.498 "is_configured": true, 00:15:36.498 "data_offset": 256, 00:15:36.498 "data_size": 7936 00:15:36.498 } 00:15:36.498 ] 00:15:36.498 } 00:15:36.498 } 00:15:36.498 }' 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:36.498 pt2' 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:36.498 19:05:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:36.498 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:15:36.498 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:36.498 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.498 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.498 [2024-12-05 19:05:54.012262] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:36.498 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.498 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=684ac437-bf68-4c71-8f0e-006d4fccec01 00:15:36.498 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@436 -- # '[' -z 684ac437-bf68-4c71-8f0e-006d4fccec01 ']' 00:15:36.498 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:36.498 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.498 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.498 [2024-12-05 19:05:54.040068] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:36.498 [2024-12-05 19:05:54.040143] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:36.498 [2024-12-05 19:05:54.040236] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:36.498 [2024-12-05 19:05:54.040323] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:36.498 [2024-12-05 19:05:54.040372] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:36.498 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.498 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:15:36.498 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.498 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.498 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@652 -- # local es=0 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.758 [2024-12-05 19:05:54.168059] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:36.758 [2024-12-05 19:05:54.169902] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:36.758 [2024-12-05 19:05:54.169981] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:36.758 [2024-12-05 19:05:54.170024] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:36.758 [2024-12-05 19:05:54.170040] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:36.758 [2024-12-05 19:05:54.170053] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:15:36.758 request: 00:15:36.758 { 00:15:36.758 "name": "raid_bdev1", 00:15:36.758 "raid_level": "raid1", 00:15:36.758 "base_bdevs": [ 00:15:36.758 "malloc1", 00:15:36.758 "malloc2" 00:15:36.758 ], 00:15:36.758 "superblock": false, 00:15:36.758 "method": "bdev_raid_create", 00:15:36.758 "req_id": 1 00:15:36.758 } 00:15:36.758 Got JSON-RPC error response 00:15:36.758 response: 00:15:36.758 { 00:15:36.758 "code": -17, 00:15:36.758 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:36.758 } 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@655 -- # es=1 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.758 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.758 [2024-12-05 19:05:54.235906] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:36.758 [2024-12-05 19:05:54.236029] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:36.758 [2024-12-05 19:05:54.236065] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:36.758 [2024-12-05 19:05:54.236106] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:36.758 [2024-12-05 19:05:54.238134] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:36.758 [2024-12-05 19:05:54.238198] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:36.758 [2024-12-05 19:05:54.238290] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:36.758 [2024-12-05 19:05:54.238335] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:36.758 pt1 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:36.759 "name": "raid_bdev1", 00:15:36.759 "uuid": "684ac437-bf68-4c71-8f0e-006d4fccec01", 00:15:36.759 "strip_size_kb": 0, 00:15:36.759 "state": "configuring", 00:15:36.759 "raid_level": "raid1", 00:15:36.759 "superblock": true, 00:15:36.759 "num_base_bdevs": 2, 00:15:36.759 "num_base_bdevs_discovered": 1, 00:15:36.759 "num_base_bdevs_operational": 2, 00:15:36.759 "base_bdevs_list": [ 00:15:36.759 { 00:15:36.759 "name": "pt1", 00:15:36.759 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:36.759 "is_configured": true, 00:15:36.759 "data_offset": 256, 00:15:36.759 "data_size": 7936 00:15:36.759 }, 00:15:36.759 { 00:15:36.759 "name": null, 00:15:36.759 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:36.759 "is_configured": false, 00:15:36.759 "data_offset": 256, 00:15:36.759 "data_size": 7936 00:15:36.759 } 00:15:36.759 ] 00:15:36.759 }' 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:36.759 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.329 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:15:37.329 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:15:37.329 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:37.329 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:37.329 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.329 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.329 [2024-12-05 19:05:54.679169] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:37.329 [2024-12-05 19:05:54.679221] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:37.329 [2024-12-05 19:05:54.679240] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:15:37.330 [2024-12-05 19:05:54.679248] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:37.330 [2024-12-05 19:05:54.679582] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:37.330 [2024-12-05 19:05:54.679598] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:37.330 [2024-12-05 19:05:54.679656] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:37.330 [2024-12-05 19:05:54.679693] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:37.330 [2024-12-05 19:05:54.679784] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:37.330 [2024-12-05 19:05:54.679792] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:37.330 [2024-12-05 19:05:54.680051] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:15:37.330 [2024-12-05 19:05:54.680171] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:37.330 [2024-12-05 19:05:54.680184] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:15:37.330 [2024-12-05 19:05:54.680280] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:37.330 pt2 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:37.330 "name": "raid_bdev1", 00:15:37.330 "uuid": "684ac437-bf68-4c71-8f0e-006d4fccec01", 00:15:37.330 "strip_size_kb": 0, 00:15:37.330 "state": "online", 00:15:37.330 "raid_level": "raid1", 00:15:37.330 "superblock": true, 00:15:37.330 "num_base_bdevs": 2, 00:15:37.330 "num_base_bdevs_discovered": 2, 00:15:37.330 "num_base_bdevs_operational": 2, 00:15:37.330 "base_bdevs_list": [ 00:15:37.330 { 00:15:37.330 "name": "pt1", 00:15:37.330 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:37.330 "is_configured": true, 00:15:37.330 "data_offset": 256, 00:15:37.330 "data_size": 7936 00:15:37.330 }, 00:15:37.330 { 00:15:37.330 "name": "pt2", 00:15:37.330 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:37.330 "is_configured": true, 00:15:37.330 "data_offset": 256, 00:15:37.330 "data_size": 7936 00:15:37.330 } 00:15:37.330 ] 00:15:37.330 }' 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:37.330 19:05:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.591 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:15:37.591 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:37.591 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:37.591 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:37.591 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:37.591 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:37.851 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:37.851 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:37.851 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.851 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.851 [2024-12-05 19:05:55.154592] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:37.851 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.851 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:37.851 "name": "raid_bdev1", 00:15:37.851 "aliases": [ 00:15:37.851 "684ac437-bf68-4c71-8f0e-006d4fccec01" 00:15:37.851 ], 00:15:37.851 "product_name": "Raid Volume", 00:15:37.851 "block_size": 4096, 00:15:37.851 "num_blocks": 7936, 00:15:37.851 "uuid": "684ac437-bf68-4c71-8f0e-006d4fccec01", 00:15:37.851 "assigned_rate_limits": { 00:15:37.851 "rw_ios_per_sec": 0, 00:15:37.851 "rw_mbytes_per_sec": 0, 00:15:37.851 "r_mbytes_per_sec": 0, 00:15:37.851 "w_mbytes_per_sec": 0 00:15:37.851 }, 00:15:37.851 "claimed": false, 00:15:37.851 "zoned": false, 00:15:37.851 "supported_io_types": { 00:15:37.851 "read": true, 00:15:37.851 "write": true, 00:15:37.851 "unmap": false, 00:15:37.851 "flush": false, 00:15:37.851 "reset": true, 00:15:37.851 "nvme_admin": false, 00:15:37.851 "nvme_io": false, 00:15:37.851 "nvme_io_md": false, 00:15:37.851 "write_zeroes": true, 00:15:37.851 "zcopy": false, 00:15:37.851 "get_zone_info": false, 00:15:37.851 "zone_management": false, 00:15:37.851 "zone_append": false, 00:15:37.851 "compare": false, 00:15:37.851 "compare_and_write": false, 00:15:37.851 "abort": false, 00:15:37.851 "seek_hole": false, 00:15:37.851 "seek_data": false, 00:15:37.851 "copy": false, 00:15:37.851 "nvme_iov_md": false 00:15:37.851 }, 00:15:37.851 "memory_domains": [ 00:15:37.851 { 00:15:37.851 "dma_device_id": "system", 00:15:37.851 "dma_device_type": 1 00:15:37.851 }, 00:15:37.851 { 00:15:37.851 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:37.851 "dma_device_type": 2 00:15:37.851 }, 00:15:37.851 { 00:15:37.851 "dma_device_id": "system", 00:15:37.851 "dma_device_type": 1 00:15:37.851 }, 00:15:37.851 { 00:15:37.851 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:37.851 "dma_device_type": 2 00:15:37.851 } 00:15:37.851 ], 00:15:37.851 "driver_specific": { 00:15:37.851 "raid": { 00:15:37.851 "uuid": "684ac437-bf68-4c71-8f0e-006d4fccec01", 00:15:37.851 "strip_size_kb": 0, 00:15:37.851 "state": "online", 00:15:37.851 "raid_level": "raid1", 00:15:37.851 "superblock": true, 00:15:37.851 "num_base_bdevs": 2, 00:15:37.851 "num_base_bdevs_discovered": 2, 00:15:37.851 "num_base_bdevs_operational": 2, 00:15:37.851 "base_bdevs_list": [ 00:15:37.851 { 00:15:37.851 "name": "pt1", 00:15:37.852 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:37.852 "is_configured": true, 00:15:37.852 "data_offset": 256, 00:15:37.852 "data_size": 7936 00:15:37.852 }, 00:15:37.852 { 00:15:37.852 "name": "pt2", 00:15:37.852 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:37.852 "is_configured": true, 00:15:37.852 "data_offset": 256, 00:15:37.852 "data_size": 7936 00:15:37.852 } 00:15:37.852 ] 00:15:37.852 } 00:15:37.852 } 00:15:37.852 }' 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:37.852 pt2' 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.852 [2024-12-05 19:05:55.386183] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:37.852 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # '[' 684ac437-bf68-4c71-8f0e-006d4fccec01 '!=' 684ac437-bf68-4c71-8f0e-006d4fccec01 ']' 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.112 [2024-12-05 19:05:55.429918] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:38.112 "name": "raid_bdev1", 00:15:38.112 "uuid": "684ac437-bf68-4c71-8f0e-006d4fccec01", 00:15:38.112 "strip_size_kb": 0, 00:15:38.112 "state": "online", 00:15:38.112 "raid_level": "raid1", 00:15:38.112 "superblock": true, 00:15:38.112 "num_base_bdevs": 2, 00:15:38.112 "num_base_bdevs_discovered": 1, 00:15:38.112 "num_base_bdevs_operational": 1, 00:15:38.112 "base_bdevs_list": [ 00:15:38.112 { 00:15:38.112 "name": null, 00:15:38.112 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.112 "is_configured": false, 00:15:38.112 "data_offset": 0, 00:15:38.112 "data_size": 7936 00:15:38.112 }, 00:15:38.112 { 00:15:38.112 "name": "pt2", 00:15:38.112 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:38.112 "is_configured": true, 00:15:38.112 "data_offset": 256, 00:15:38.112 "data_size": 7936 00:15:38.112 } 00:15:38.112 ] 00:15:38.112 }' 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:38.112 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.372 [2024-12-05 19:05:55.821227] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:38.372 [2024-12-05 19:05:55.821305] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:38.372 [2024-12-05 19:05:55.821403] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:38.372 [2024-12-05 19:05:55.821461] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:38.372 [2024-12-05 19:05:55.821492] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@519 -- # i=1 00:15:38.372 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.373 [2024-12-05 19:05:55.897109] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:38.373 [2024-12-05 19:05:55.897225] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:38.373 [2024-12-05 19:05:55.897258] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:15:38.373 [2024-12-05 19:05:55.897283] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:38.373 [2024-12-05 19:05:55.899377] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:38.373 [2024-12-05 19:05:55.899447] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:38.373 [2024-12-05 19:05:55.899529] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:38.373 [2024-12-05 19:05:55.899590] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:38.373 [2024-12-05 19:05:55.899690] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:15:38.373 [2024-12-05 19:05:55.899716] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:38.373 [2024-12-05 19:05:55.899988] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:15:38.373 [2024-12-05 19:05:55.900135] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:15:38.373 [2024-12-05 19:05:55.900177] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:15:38.373 [2024-12-05 19:05:55.900303] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:38.373 pt2 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.373 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.633 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:38.633 "name": "raid_bdev1", 00:15:38.633 "uuid": "684ac437-bf68-4c71-8f0e-006d4fccec01", 00:15:38.633 "strip_size_kb": 0, 00:15:38.633 "state": "online", 00:15:38.633 "raid_level": "raid1", 00:15:38.633 "superblock": true, 00:15:38.633 "num_base_bdevs": 2, 00:15:38.633 "num_base_bdevs_discovered": 1, 00:15:38.633 "num_base_bdevs_operational": 1, 00:15:38.633 "base_bdevs_list": [ 00:15:38.633 { 00:15:38.633 "name": null, 00:15:38.633 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.633 "is_configured": false, 00:15:38.633 "data_offset": 256, 00:15:38.633 "data_size": 7936 00:15:38.633 }, 00:15:38.633 { 00:15:38.633 "name": "pt2", 00:15:38.633 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:38.633 "is_configured": true, 00:15:38.633 "data_offset": 256, 00:15:38.633 "data_size": 7936 00:15:38.633 } 00:15:38.633 ] 00:15:38.633 }' 00:15:38.633 19:05:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:38.633 19:05:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.893 [2024-12-05 19:05:56.356332] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:38.893 [2024-12-05 19:05:56.356401] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:38.893 [2024-12-05 19:05:56.356452] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:38.893 [2024-12-05 19:05:56.356486] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:38.893 [2024-12-05 19:05:56.356496] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.893 [2024-12-05 19:05:56.420221] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:38.893 [2024-12-05 19:05:56.420318] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:38.893 [2024-12-05 19:05:56.420349] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:15:38.893 [2024-12-05 19:05:56.420378] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:38.893 [2024-12-05 19:05:56.422411] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:38.893 [2024-12-05 19:05:56.422481] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:38.893 [2024-12-05 19:05:56.422557] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:38.893 [2024-12-05 19:05:56.422622] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:38.893 [2024-12-05 19:05:56.422731] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:15:38.893 [2024-12-05 19:05:56.422791] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:38.893 [2024-12-05 19:05:56.422840] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:15:38.893 [2024-12-05 19:05:56.422914] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:38.893 [2024-12-05 19:05:56.423027] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:15:38.893 [2024-12-05 19:05:56.423067] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:38.893 [2024-12-05 19:05:56.423291] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:38.893 [2024-12-05 19:05:56.423429] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:15:38.893 [2024-12-05 19:05:56.423467] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:15:38.893 [2024-12-05 19:05:56.423599] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:38.893 pt1 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.893 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.154 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:39.154 "name": "raid_bdev1", 00:15:39.154 "uuid": "684ac437-bf68-4c71-8f0e-006d4fccec01", 00:15:39.154 "strip_size_kb": 0, 00:15:39.154 "state": "online", 00:15:39.154 "raid_level": "raid1", 00:15:39.154 "superblock": true, 00:15:39.154 "num_base_bdevs": 2, 00:15:39.154 "num_base_bdevs_discovered": 1, 00:15:39.154 "num_base_bdevs_operational": 1, 00:15:39.154 "base_bdevs_list": [ 00:15:39.154 { 00:15:39.154 "name": null, 00:15:39.154 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:39.154 "is_configured": false, 00:15:39.154 "data_offset": 256, 00:15:39.154 "data_size": 7936 00:15:39.154 }, 00:15:39.154 { 00:15:39.154 "name": "pt2", 00:15:39.154 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:39.154 "is_configured": true, 00:15:39.154 "data_offset": 256, 00:15:39.154 "data_size": 7936 00:15:39.154 } 00:15:39.154 ] 00:15:39.154 }' 00:15:39.154 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:39.154 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.414 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:15:39.414 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:39.414 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.414 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.414 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.414 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:15:39.414 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:39.414 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:39.415 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:15:39.415 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.415 [2024-12-05 19:05:56.931782] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:39.415 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:39.415 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # '[' 684ac437-bf68-4c71-8f0e-006d4fccec01 '!=' 684ac437-bf68-4c71-8f0e-006d4fccec01 ']' 00:15:39.415 19:05:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@563 -- # killprocess 93714 00:15:39.415 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@954 -- # '[' -z 93714 ']' 00:15:39.415 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@958 -- # kill -0 93714 00:15:39.675 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@959 -- # uname 00:15:39.675 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:39.675 19:05:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 93714 00:15:39.675 killing process with pid 93714 00:15:39.675 19:05:57 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:39.675 19:05:57 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:39.675 19:05:57 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 93714' 00:15:39.675 19:05:57 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@973 -- # kill 93714 00:15:39.675 [2024-12-05 19:05:57.013699] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:39.675 [2024-12-05 19:05:57.013761] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:39.675 [2024-12-05 19:05:57.013801] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:39.675 [2024-12-05 19:05:57.013809] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:15:39.675 19:05:57 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@978 -- # wait 93714 00:15:39.675 [2024-12-05 19:05:57.035887] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:39.935 19:05:57 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@565 -- # return 0 00:15:39.935 00:15:39.935 real 0m4.955s 00:15:39.935 user 0m8.117s 00:15:39.935 sys 0m1.113s 00:15:39.935 19:05:57 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:39.935 ************************************ 00:15:39.935 END TEST raid_superblock_test_4k 00:15:39.935 ************************************ 00:15:39.935 19:05:57 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.935 19:05:57 bdev_raid -- bdev/bdev_raid.sh@999 -- # '[' true = true ']' 00:15:39.935 19:05:57 bdev_raid -- bdev/bdev_raid.sh@1000 -- # run_test raid_rebuild_test_sb_4k raid_rebuild_test raid1 2 true false true 00:15:39.935 19:05:57 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:15:39.935 19:05:57 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:39.935 19:05:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:39.935 ************************************ 00:15:39.935 START TEST raid_rebuild_test_sb_4k 00:15:39.935 ************************************ 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:39.935 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:39.936 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@597 -- # raid_pid=93996 00:15:39.936 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@598 -- # waitforlisten 93996 00:15:39.936 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:39.936 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@835 -- # '[' -z 93996 ']' 00:15:39.936 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:39.936 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:39.936 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:39.936 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:39.936 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:39.936 19:05:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.936 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:39.936 Zero copy mechanism will not be used. 00:15:39.936 [2024-12-05 19:05:57.439923] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:15:39.936 [2024-12-05 19:05:57.440070] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid93996 ] 00:15:40.196 [2024-12-05 19:05:57.597513] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:40.196 [2024-12-05 19:05:57.623075] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:40.196 [2024-12-05 19:05:57.666696] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:40.196 [2024-12-05 19:05:57.666725] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # return 0 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1_malloc 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:40.766 BaseBdev1_malloc 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:40.766 [2024-12-05 19:05:58.262551] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:40.766 [2024-12-05 19:05:58.262620] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:40.766 [2024-12-05 19:05:58.262666] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:40.766 [2024-12-05 19:05:58.262693] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:40.766 [2024-12-05 19:05:58.264803] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:40.766 [2024-12-05 19:05:58.264839] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:40.766 BaseBdev1 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2_malloc 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:40.766 BaseBdev2_malloc 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:40.766 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.767 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:40.767 [2024-12-05 19:05:58.291000] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:40.767 [2024-12-05 19:05:58.291056] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:40.767 [2024-12-05 19:05:58.291094] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:40.767 [2024-12-05 19:05:58.291103] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:40.767 [2024-12-05 19:05:58.293130] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:40.767 [2024-12-05 19:05:58.293173] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:40.767 BaseBdev2 00:15:40.767 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.767 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -b spare_malloc 00:15:40.767 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.767 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:40.767 spare_malloc 00:15:40.767 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:40.767 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:40.767 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:40.767 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.027 spare_delay 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.027 [2024-12-05 19:05:58.331415] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:41.027 [2024-12-05 19:05:58.331554] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:41.027 [2024-12-05 19:05:58.331576] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:15:41.027 [2024-12-05 19:05:58.331585] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:41.027 [2024-12-05 19:05:58.333630] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:41.027 [2024-12-05 19:05:58.333678] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:41.027 spare 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.027 [2024-12-05 19:05:58.343432] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:41.027 [2024-12-05 19:05:58.345172] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:41.027 [2024-12-05 19:05:58.345323] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:41.027 [2024-12-05 19:05:58.345335] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:41.027 [2024-12-05 19:05:58.345595] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:15:41.027 [2024-12-05 19:05:58.345743] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:41.027 [2024-12-05 19:05:58.345756] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:41.027 [2024-12-05 19:05:58.345875] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.027 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:41.028 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.028 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.028 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.028 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:41.028 "name": "raid_bdev1", 00:15:41.028 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:41.028 "strip_size_kb": 0, 00:15:41.028 "state": "online", 00:15:41.028 "raid_level": "raid1", 00:15:41.028 "superblock": true, 00:15:41.028 "num_base_bdevs": 2, 00:15:41.028 "num_base_bdevs_discovered": 2, 00:15:41.028 "num_base_bdevs_operational": 2, 00:15:41.028 "base_bdevs_list": [ 00:15:41.028 { 00:15:41.028 "name": "BaseBdev1", 00:15:41.028 "uuid": "8dccce37-ddd5-500a-b5ff-86370447a150", 00:15:41.028 "is_configured": true, 00:15:41.028 "data_offset": 256, 00:15:41.028 "data_size": 7936 00:15:41.028 }, 00:15:41.028 { 00:15:41.028 "name": "BaseBdev2", 00:15:41.028 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:41.028 "is_configured": true, 00:15:41.028 "data_offset": 256, 00:15:41.028 "data_size": 7936 00:15:41.028 } 00:15:41.028 ] 00:15:41.028 }' 00:15:41.028 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:41.028 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.296 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:41.296 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:41.296 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.296 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.296 [2024-12-05 19:05:58.806846] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:41.296 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.296 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:15:41.296 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:41.296 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.296 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:41.296 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.593 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:41.593 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:15:41.593 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:41.593 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:41.593 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:41.593 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:41.593 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:41.593 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:41.593 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:41.593 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:41.593 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:41.593 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:15:41.593 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:41.593 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:41.593 19:05:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:41.593 [2024-12-05 19:05:59.058192] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:41.593 /dev/nbd0 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:41.593 1+0 records in 00:15:41.593 1+0 records out 00:15:41.593 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000308456 s, 13.3 MB/s 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:15:41.593 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:15:42.179 7936+0 records in 00:15:42.179 7936+0 records out 00:15:42.179 32505856 bytes (33 MB, 31 MiB) copied, 0.590495 s, 55.0 MB/s 00:15:42.179 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:42.179 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:42.179 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:42.179 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:42.179 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:15:42.179 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:42.179 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:42.440 [2024-12-05 19:05:59.928477] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:42.440 [2024-12-05 19:05:59.944560] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:42.440 19:05:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.700 19:06:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:42.700 "name": "raid_bdev1", 00:15:42.700 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:42.700 "strip_size_kb": 0, 00:15:42.700 "state": "online", 00:15:42.700 "raid_level": "raid1", 00:15:42.700 "superblock": true, 00:15:42.700 "num_base_bdevs": 2, 00:15:42.700 "num_base_bdevs_discovered": 1, 00:15:42.700 "num_base_bdevs_operational": 1, 00:15:42.700 "base_bdevs_list": [ 00:15:42.700 { 00:15:42.700 "name": null, 00:15:42.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.700 "is_configured": false, 00:15:42.700 "data_offset": 0, 00:15:42.700 "data_size": 7936 00:15:42.700 }, 00:15:42.700 { 00:15:42.700 "name": "BaseBdev2", 00:15:42.700 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:42.700 "is_configured": true, 00:15:42.700 "data_offset": 256, 00:15:42.700 "data_size": 7936 00:15:42.700 } 00:15:42.700 ] 00:15:42.700 }' 00:15:42.700 19:06:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:42.700 19:06:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:42.960 19:06:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:42.960 19:06:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:42.960 19:06:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:42.960 [2024-12-05 19:06:00.391914] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:42.960 [2024-12-05 19:06:00.396956] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019c960 00:15:42.960 19:06:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:42.960 19:06:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:42.960 [2024-12-05 19:06:00.398884] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:43.902 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:43.902 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:43.902 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:43.902 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:43.902 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:43.902 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:43.902 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:43.902 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:43.902 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:43.902 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:43.902 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:43.902 "name": "raid_bdev1", 00:15:43.902 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:43.902 "strip_size_kb": 0, 00:15:43.902 "state": "online", 00:15:43.902 "raid_level": "raid1", 00:15:43.902 "superblock": true, 00:15:43.902 "num_base_bdevs": 2, 00:15:43.902 "num_base_bdevs_discovered": 2, 00:15:43.902 "num_base_bdevs_operational": 2, 00:15:43.902 "process": { 00:15:43.902 "type": "rebuild", 00:15:43.902 "target": "spare", 00:15:43.902 "progress": { 00:15:43.902 "blocks": 2560, 00:15:43.902 "percent": 32 00:15:43.902 } 00:15:43.902 }, 00:15:43.902 "base_bdevs_list": [ 00:15:43.902 { 00:15:43.902 "name": "spare", 00:15:43.902 "uuid": "8d63ffc5-2c94-589b-8402-381e1b25ebc3", 00:15:43.902 "is_configured": true, 00:15:43.902 "data_offset": 256, 00:15:43.902 "data_size": 7936 00:15:43.902 }, 00:15:43.902 { 00:15:43.902 "name": "BaseBdev2", 00:15:43.902 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:43.902 "is_configured": true, 00:15:43.902 "data_offset": 256, 00:15:43.902 "data_size": 7936 00:15:43.902 } 00:15:43.902 ] 00:15:43.902 }' 00:15:43.902 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.162 [2024-12-05 19:06:01.563353] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:44.162 [2024-12-05 19:06:01.603263] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:44.162 [2024-12-05 19:06:01.603316] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:44.162 [2024-12-05 19:06:01.603334] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:44.162 [2024-12-05 19:06:01.603341] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.162 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:44.162 "name": "raid_bdev1", 00:15:44.162 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:44.162 "strip_size_kb": 0, 00:15:44.162 "state": "online", 00:15:44.162 "raid_level": "raid1", 00:15:44.162 "superblock": true, 00:15:44.162 "num_base_bdevs": 2, 00:15:44.162 "num_base_bdevs_discovered": 1, 00:15:44.162 "num_base_bdevs_operational": 1, 00:15:44.162 "base_bdevs_list": [ 00:15:44.162 { 00:15:44.162 "name": null, 00:15:44.162 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:44.162 "is_configured": false, 00:15:44.162 "data_offset": 0, 00:15:44.162 "data_size": 7936 00:15:44.162 }, 00:15:44.162 { 00:15:44.162 "name": "BaseBdev2", 00:15:44.162 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:44.162 "is_configured": true, 00:15:44.162 "data_offset": 256, 00:15:44.163 "data_size": 7936 00:15:44.163 } 00:15:44.163 ] 00:15:44.163 }' 00:15:44.163 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:44.163 19:06:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:44.732 "name": "raid_bdev1", 00:15:44.732 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:44.732 "strip_size_kb": 0, 00:15:44.732 "state": "online", 00:15:44.732 "raid_level": "raid1", 00:15:44.732 "superblock": true, 00:15:44.732 "num_base_bdevs": 2, 00:15:44.732 "num_base_bdevs_discovered": 1, 00:15:44.732 "num_base_bdevs_operational": 1, 00:15:44.732 "base_bdevs_list": [ 00:15:44.732 { 00:15:44.732 "name": null, 00:15:44.732 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:44.732 "is_configured": false, 00:15:44.732 "data_offset": 0, 00:15:44.732 "data_size": 7936 00:15:44.732 }, 00:15:44.732 { 00:15:44.732 "name": "BaseBdev2", 00:15:44.732 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:44.732 "is_configured": true, 00:15:44.732 "data_offset": 256, 00:15:44.732 "data_size": 7936 00:15:44.732 } 00:15:44.732 ] 00:15:44.732 }' 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.732 [2024-12-05 19:06:02.238958] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:44.732 [2024-12-05 19:06:02.243706] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019ca30 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:44.732 19:06:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:44.732 [2024-12-05 19:06:02.245604] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:46.113 "name": "raid_bdev1", 00:15:46.113 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:46.113 "strip_size_kb": 0, 00:15:46.113 "state": "online", 00:15:46.113 "raid_level": "raid1", 00:15:46.113 "superblock": true, 00:15:46.113 "num_base_bdevs": 2, 00:15:46.113 "num_base_bdevs_discovered": 2, 00:15:46.113 "num_base_bdevs_operational": 2, 00:15:46.113 "process": { 00:15:46.113 "type": "rebuild", 00:15:46.113 "target": "spare", 00:15:46.113 "progress": { 00:15:46.113 "blocks": 2560, 00:15:46.113 "percent": 32 00:15:46.113 } 00:15:46.113 }, 00:15:46.113 "base_bdevs_list": [ 00:15:46.113 { 00:15:46.113 "name": "spare", 00:15:46.113 "uuid": "8d63ffc5-2c94-589b-8402-381e1b25ebc3", 00:15:46.113 "is_configured": true, 00:15:46.113 "data_offset": 256, 00:15:46.113 "data_size": 7936 00:15:46.113 }, 00:15:46.113 { 00:15:46.113 "name": "BaseBdev2", 00:15:46.113 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:46.113 "is_configured": true, 00:15:46.113 "data_offset": 256, 00:15:46.113 "data_size": 7936 00:15:46.113 } 00:15:46.113 ] 00:15:46.113 }' 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:15:46.113 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@706 -- # local timeout=553 00:15:46.113 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:46.114 "name": "raid_bdev1", 00:15:46.114 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:46.114 "strip_size_kb": 0, 00:15:46.114 "state": "online", 00:15:46.114 "raid_level": "raid1", 00:15:46.114 "superblock": true, 00:15:46.114 "num_base_bdevs": 2, 00:15:46.114 "num_base_bdevs_discovered": 2, 00:15:46.114 "num_base_bdevs_operational": 2, 00:15:46.114 "process": { 00:15:46.114 "type": "rebuild", 00:15:46.114 "target": "spare", 00:15:46.114 "progress": { 00:15:46.114 "blocks": 2816, 00:15:46.114 "percent": 35 00:15:46.114 } 00:15:46.114 }, 00:15:46.114 "base_bdevs_list": [ 00:15:46.114 { 00:15:46.114 "name": "spare", 00:15:46.114 "uuid": "8d63ffc5-2c94-589b-8402-381e1b25ebc3", 00:15:46.114 "is_configured": true, 00:15:46.114 "data_offset": 256, 00:15:46.114 "data_size": 7936 00:15:46.114 }, 00:15:46.114 { 00:15:46.114 "name": "BaseBdev2", 00:15:46.114 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:46.114 "is_configured": true, 00:15:46.114 "data_offset": 256, 00:15:46.114 "data_size": 7936 00:15:46.114 } 00:15:46.114 ] 00:15:46.114 }' 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:46.114 19:06:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:47.052 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:47.052 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:47.052 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:47.052 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:47.052 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:47.052 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:47.052 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:47.052 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:47.052 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:47.052 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:47.052 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:47.052 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:47.052 "name": "raid_bdev1", 00:15:47.052 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:47.052 "strip_size_kb": 0, 00:15:47.052 "state": "online", 00:15:47.052 "raid_level": "raid1", 00:15:47.052 "superblock": true, 00:15:47.052 "num_base_bdevs": 2, 00:15:47.052 "num_base_bdevs_discovered": 2, 00:15:47.052 "num_base_bdevs_operational": 2, 00:15:47.052 "process": { 00:15:47.052 "type": "rebuild", 00:15:47.052 "target": "spare", 00:15:47.052 "progress": { 00:15:47.052 "blocks": 5632, 00:15:47.052 "percent": 70 00:15:47.052 } 00:15:47.053 }, 00:15:47.053 "base_bdevs_list": [ 00:15:47.053 { 00:15:47.053 "name": "spare", 00:15:47.053 "uuid": "8d63ffc5-2c94-589b-8402-381e1b25ebc3", 00:15:47.053 "is_configured": true, 00:15:47.053 "data_offset": 256, 00:15:47.053 "data_size": 7936 00:15:47.053 }, 00:15:47.053 { 00:15:47.053 "name": "BaseBdev2", 00:15:47.053 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:47.053 "is_configured": true, 00:15:47.053 "data_offset": 256, 00:15:47.053 "data_size": 7936 00:15:47.053 } 00:15:47.053 ] 00:15:47.053 }' 00:15:47.053 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:47.312 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:47.312 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:47.312 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:47.312 19:06:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:47.880 [2024-12-05 19:06:05.355584] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:47.880 [2024-12-05 19:06:05.355683] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:47.880 [2024-12-05 19:06:05.355789] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:48.139 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:48.139 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:48.139 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:48.139 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:48.139 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:48.139 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:48.139 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.139 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:48.139 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.139 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.139 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:48.398 "name": "raid_bdev1", 00:15:48.398 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:48.398 "strip_size_kb": 0, 00:15:48.398 "state": "online", 00:15:48.398 "raid_level": "raid1", 00:15:48.398 "superblock": true, 00:15:48.398 "num_base_bdevs": 2, 00:15:48.398 "num_base_bdevs_discovered": 2, 00:15:48.398 "num_base_bdevs_operational": 2, 00:15:48.398 "base_bdevs_list": [ 00:15:48.398 { 00:15:48.398 "name": "spare", 00:15:48.398 "uuid": "8d63ffc5-2c94-589b-8402-381e1b25ebc3", 00:15:48.398 "is_configured": true, 00:15:48.398 "data_offset": 256, 00:15:48.398 "data_size": 7936 00:15:48.398 }, 00:15:48.398 { 00:15:48.398 "name": "BaseBdev2", 00:15:48.398 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:48.398 "is_configured": true, 00:15:48.398 "data_offset": 256, 00:15:48.398 "data_size": 7936 00:15:48.398 } 00:15:48.398 ] 00:15:48.398 }' 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@709 -- # break 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:48.398 "name": "raid_bdev1", 00:15:48.398 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:48.398 "strip_size_kb": 0, 00:15:48.398 "state": "online", 00:15:48.398 "raid_level": "raid1", 00:15:48.398 "superblock": true, 00:15:48.398 "num_base_bdevs": 2, 00:15:48.398 "num_base_bdevs_discovered": 2, 00:15:48.398 "num_base_bdevs_operational": 2, 00:15:48.398 "base_bdevs_list": [ 00:15:48.398 { 00:15:48.398 "name": "spare", 00:15:48.398 "uuid": "8d63ffc5-2c94-589b-8402-381e1b25ebc3", 00:15:48.398 "is_configured": true, 00:15:48.398 "data_offset": 256, 00:15:48.398 "data_size": 7936 00:15:48.398 }, 00:15:48.398 { 00:15:48.398 "name": "BaseBdev2", 00:15:48.398 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:48.398 "is_configured": true, 00:15:48.398 "data_offset": 256, 00:15:48.398 "data_size": 7936 00:15:48.398 } 00:15:48.398 ] 00:15:48.398 }' 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:48.398 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:48.657 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.657 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.657 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:48.657 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.657 19:06:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.657 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:48.657 "name": "raid_bdev1", 00:15:48.657 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:48.657 "strip_size_kb": 0, 00:15:48.657 "state": "online", 00:15:48.657 "raid_level": "raid1", 00:15:48.657 "superblock": true, 00:15:48.657 "num_base_bdevs": 2, 00:15:48.657 "num_base_bdevs_discovered": 2, 00:15:48.657 "num_base_bdevs_operational": 2, 00:15:48.657 "base_bdevs_list": [ 00:15:48.657 { 00:15:48.657 "name": "spare", 00:15:48.657 "uuid": "8d63ffc5-2c94-589b-8402-381e1b25ebc3", 00:15:48.657 "is_configured": true, 00:15:48.657 "data_offset": 256, 00:15:48.657 "data_size": 7936 00:15:48.657 }, 00:15:48.657 { 00:15:48.657 "name": "BaseBdev2", 00:15:48.657 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:48.657 "is_configured": true, 00:15:48.657 "data_offset": 256, 00:15:48.657 "data_size": 7936 00:15:48.657 } 00:15:48.657 ] 00:15:48.657 }' 00:15:48.657 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:48.657 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.917 [2024-12-05 19:06:06.406126] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:48.917 [2024-12-05 19:06:06.406205] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:48.917 [2024-12-05 19:06:06.406295] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:48.917 [2024-12-05 19:06:06.406374] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:48.917 [2024-12-05 19:06:06.406386] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # jq length 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:48.917 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:49.176 /dev/nbd0 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:49.176 1+0 records in 00:15:49.176 1+0 records out 00:15:49.176 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000491118 s, 8.3 MB/s 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:49.176 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:49.436 /dev/nbd1 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # local i 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@877 -- # break 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:49.436 1+0 records in 00:15:49.436 1+0 records out 00:15:49.436 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000358479 s, 11.4 MB/s 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@890 -- # size=4096 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@893 -- # return 0 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:49.436 19:06:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:15:49.695 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:49.695 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:49.695 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:49.695 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:49.695 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:15:49.695 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:49.695 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:49.695 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:49.695 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:49.695 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:49.695 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:49.695 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:49.695 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:49.955 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.214 [2024-12-05 19:06:07.515645] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:50.214 [2024-12-05 19:06:07.515761] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:50.214 [2024-12-05 19:06:07.515814] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:15:50.214 [2024-12-05 19:06:07.515847] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:50.214 [2024-12-05 19:06:07.517967] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:50.214 [2024-12-05 19:06:07.518044] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:50.214 [2024-12-05 19:06:07.518126] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:50.214 [2024-12-05 19:06:07.518180] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:50.214 [2024-12-05 19:06:07.518296] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:50.214 spare 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.214 [2024-12-05 19:06:07.618187] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:15:50.214 [2024-12-05 19:06:07.618210] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:50.214 [2024-12-05 19:06:07.618447] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb1b0 00:15:50.214 [2024-12-05 19:06:07.618576] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:15:50.214 [2024-12-05 19:06:07.618595] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:15:50.214 [2024-12-05 19:06:07.618810] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:50.214 "name": "raid_bdev1", 00:15:50.214 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:50.214 "strip_size_kb": 0, 00:15:50.214 "state": "online", 00:15:50.214 "raid_level": "raid1", 00:15:50.214 "superblock": true, 00:15:50.214 "num_base_bdevs": 2, 00:15:50.214 "num_base_bdevs_discovered": 2, 00:15:50.214 "num_base_bdevs_operational": 2, 00:15:50.214 "base_bdevs_list": [ 00:15:50.214 { 00:15:50.214 "name": "spare", 00:15:50.214 "uuid": "8d63ffc5-2c94-589b-8402-381e1b25ebc3", 00:15:50.214 "is_configured": true, 00:15:50.214 "data_offset": 256, 00:15:50.214 "data_size": 7936 00:15:50.214 }, 00:15:50.214 { 00:15:50.214 "name": "BaseBdev2", 00:15:50.214 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:50.214 "is_configured": true, 00:15:50.214 "data_offset": 256, 00:15:50.214 "data_size": 7936 00:15:50.214 } 00:15:50.214 ] 00:15:50.214 }' 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:50.214 19:06:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:50.784 "name": "raid_bdev1", 00:15:50.784 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:50.784 "strip_size_kb": 0, 00:15:50.784 "state": "online", 00:15:50.784 "raid_level": "raid1", 00:15:50.784 "superblock": true, 00:15:50.784 "num_base_bdevs": 2, 00:15:50.784 "num_base_bdevs_discovered": 2, 00:15:50.784 "num_base_bdevs_operational": 2, 00:15:50.784 "base_bdevs_list": [ 00:15:50.784 { 00:15:50.784 "name": "spare", 00:15:50.784 "uuid": "8d63ffc5-2c94-589b-8402-381e1b25ebc3", 00:15:50.784 "is_configured": true, 00:15:50.784 "data_offset": 256, 00:15:50.784 "data_size": 7936 00:15:50.784 }, 00:15:50.784 { 00:15:50.784 "name": "BaseBdev2", 00:15:50.784 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:50.784 "is_configured": true, 00:15:50.784 "data_offset": 256, 00:15:50.784 "data_size": 7936 00:15:50.784 } 00:15:50.784 ] 00:15:50.784 }' 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.784 [2024-12-05 19:06:08.302335] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:50.784 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:50.785 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:50.785 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:50.785 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:50.785 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:50.785 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:50.785 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.785 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:50.785 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:50.785 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:50.785 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.044 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:51.044 "name": "raid_bdev1", 00:15:51.044 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:51.044 "strip_size_kb": 0, 00:15:51.044 "state": "online", 00:15:51.044 "raid_level": "raid1", 00:15:51.044 "superblock": true, 00:15:51.044 "num_base_bdevs": 2, 00:15:51.044 "num_base_bdevs_discovered": 1, 00:15:51.044 "num_base_bdevs_operational": 1, 00:15:51.044 "base_bdevs_list": [ 00:15:51.044 { 00:15:51.044 "name": null, 00:15:51.044 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:51.044 "is_configured": false, 00:15:51.044 "data_offset": 0, 00:15:51.044 "data_size": 7936 00:15:51.044 }, 00:15:51.044 { 00:15:51.044 "name": "BaseBdev2", 00:15:51.044 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:51.044 "is_configured": true, 00:15:51.044 "data_offset": 256, 00:15:51.044 "data_size": 7936 00:15:51.044 } 00:15:51.044 ] 00:15:51.044 }' 00:15:51.044 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:51.044 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:51.304 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:51.304 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:51.304 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:51.304 [2024-12-05 19:06:08.765537] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:51.304 [2024-12-05 19:06:08.765782] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:51.304 [2024-12-05 19:06:08.765844] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:51.304 [2024-12-05 19:06:08.765969] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:51.304 [2024-12-05 19:06:08.770867] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb280 00:15:51.304 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:51.304 19:06:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:51.304 [2024-12-05 19:06:08.772772] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:52.242 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:52.242 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:52.242 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:52.242 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:52.242 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:52.242 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.242 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:52.242 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.242 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:52.501 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:52.502 "name": "raid_bdev1", 00:15:52.502 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:52.502 "strip_size_kb": 0, 00:15:52.502 "state": "online", 00:15:52.502 "raid_level": "raid1", 00:15:52.502 "superblock": true, 00:15:52.502 "num_base_bdevs": 2, 00:15:52.502 "num_base_bdevs_discovered": 2, 00:15:52.502 "num_base_bdevs_operational": 2, 00:15:52.502 "process": { 00:15:52.502 "type": "rebuild", 00:15:52.502 "target": "spare", 00:15:52.502 "progress": { 00:15:52.502 "blocks": 2560, 00:15:52.502 "percent": 32 00:15:52.502 } 00:15:52.502 }, 00:15:52.502 "base_bdevs_list": [ 00:15:52.502 { 00:15:52.502 "name": "spare", 00:15:52.502 "uuid": "8d63ffc5-2c94-589b-8402-381e1b25ebc3", 00:15:52.502 "is_configured": true, 00:15:52.502 "data_offset": 256, 00:15:52.502 "data_size": 7936 00:15:52.502 }, 00:15:52.502 { 00:15:52.502 "name": "BaseBdev2", 00:15:52.502 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:52.502 "is_configured": true, 00:15:52.502 "data_offset": 256, 00:15:52.502 "data_size": 7936 00:15:52.502 } 00:15:52.502 ] 00:15:52.502 }' 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:52.502 [2024-12-05 19:06:09.940823] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:52.502 [2024-12-05 19:06:09.976642] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:52.502 [2024-12-05 19:06:09.976768] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:52.502 [2024-12-05 19:06:09.976809] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:52.502 [2024-12-05 19:06:09.976831] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:52.502 19:06:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:52.502 19:06:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:52.502 19:06:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:52.502 "name": "raid_bdev1", 00:15:52.502 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:52.502 "strip_size_kb": 0, 00:15:52.502 "state": "online", 00:15:52.502 "raid_level": "raid1", 00:15:52.502 "superblock": true, 00:15:52.502 "num_base_bdevs": 2, 00:15:52.502 "num_base_bdevs_discovered": 1, 00:15:52.502 "num_base_bdevs_operational": 1, 00:15:52.502 "base_bdevs_list": [ 00:15:52.502 { 00:15:52.502 "name": null, 00:15:52.502 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:52.502 "is_configured": false, 00:15:52.502 "data_offset": 0, 00:15:52.502 "data_size": 7936 00:15:52.502 }, 00:15:52.502 { 00:15:52.502 "name": "BaseBdev2", 00:15:52.502 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:52.502 "is_configured": true, 00:15:52.502 "data_offset": 256, 00:15:52.502 "data_size": 7936 00:15:52.502 } 00:15:52.502 ] 00:15:52.502 }' 00:15:52.502 19:06:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:52.502 19:06:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:53.071 19:06:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:53.072 19:06:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:53.072 19:06:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:53.072 [2024-12-05 19:06:10.436448] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:53.072 [2024-12-05 19:06:10.436556] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:53.072 [2024-12-05 19:06:10.436611] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:15:53.072 [2024-12-05 19:06:10.436638] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:53.072 [2024-12-05 19:06:10.437082] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:53.072 [2024-12-05 19:06:10.437142] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:53.072 [2024-12-05 19:06:10.437243] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:53.072 [2024-12-05 19:06:10.437282] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:53.072 [2024-12-05 19:06:10.437327] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:53.072 [2024-12-05 19:06:10.437395] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:53.072 [2024-12-05 19:06:10.441806] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb350 00:15:53.072 spare 00:15:53.072 19:06:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:53.072 19:06:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:53.072 [2024-12-05 19:06:10.443708] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:54.011 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:54.011 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:54.011 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:54.011 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:54.011 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:54.011 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.011 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.011 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:54.011 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:54.011 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.011 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:54.011 "name": "raid_bdev1", 00:15:54.011 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:54.011 "strip_size_kb": 0, 00:15:54.011 "state": "online", 00:15:54.011 "raid_level": "raid1", 00:15:54.011 "superblock": true, 00:15:54.011 "num_base_bdevs": 2, 00:15:54.011 "num_base_bdevs_discovered": 2, 00:15:54.011 "num_base_bdevs_operational": 2, 00:15:54.011 "process": { 00:15:54.011 "type": "rebuild", 00:15:54.011 "target": "spare", 00:15:54.011 "progress": { 00:15:54.011 "blocks": 2560, 00:15:54.011 "percent": 32 00:15:54.011 } 00:15:54.011 }, 00:15:54.011 "base_bdevs_list": [ 00:15:54.011 { 00:15:54.012 "name": "spare", 00:15:54.012 "uuid": "8d63ffc5-2c94-589b-8402-381e1b25ebc3", 00:15:54.012 "is_configured": true, 00:15:54.012 "data_offset": 256, 00:15:54.012 "data_size": 7936 00:15:54.012 }, 00:15:54.012 { 00:15:54.012 "name": "BaseBdev2", 00:15:54.012 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:54.012 "is_configured": true, 00:15:54.012 "data_offset": 256, 00:15:54.012 "data_size": 7936 00:15:54.012 } 00:15:54.012 ] 00:15:54.012 }' 00:15:54.012 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:54.012 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:54.012 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:54.272 [2024-12-05 19:06:11.583779] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:54.272 [2024-12-05 19:06:11.647520] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:54.272 [2024-12-05 19:06:11.647589] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:54.272 [2024-12-05 19:06:11.647603] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:54.272 [2024-12-05 19:06:11.647612] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:54.272 "name": "raid_bdev1", 00:15:54.272 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:54.272 "strip_size_kb": 0, 00:15:54.272 "state": "online", 00:15:54.272 "raid_level": "raid1", 00:15:54.272 "superblock": true, 00:15:54.272 "num_base_bdevs": 2, 00:15:54.272 "num_base_bdevs_discovered": 1, 00:15:54.272 "num_base_bdevs_operational": 1, 00:15:54.272 "base_bdevs_list": [ 00:15:54.272 { 00:15:54.272 "name": null, 00:15:54.272 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:54.272 "is_configured": false, 00:15:54.272 "data_offset": 0, 00:15:54.272 "data_size": 7936 00:15:54.272 }, 00:15:54.272 { 00:15:54.272 "name": "BaseBdev2", 00:15:54.272 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:54.272 "is_configured": true, 00:15:54.272 "data_offset": 256, 00:15:54.272 "data_size": 7936 00:15:54.272 } 00:15:54.272 ] 00:15:54.272 }' 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:54.272 19:06:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:54.843 "name": "raid_bdev1", 00:15:54.843 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:54.843 "strip_size_kb": 0, 00:15:54.843 "state": "online", 00:15:54.843 "raid_level": "raid1", 00:15:54.843 "superblock": true, 00:15:54.843 "num_base_bdevs": 2, 00:15:54.843 "num_base_bdevs_discovered": 1, 00:15:54.843 "num_base_bdevs_operational": 1, 00:15:54.843 "base_bdevs_list": [ 00:15:54.843 { 00:15:54.843 "name": null, 00:15:54.843 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:54.843 "is_configured": false, 00:15:54.843 "data_offset": 0, 00:15:54.843 "data_size": 7936 00:15:54.843 }, 00:15:54.843 { 00:15:54.843 "name": "BaseBdev2", 00:15:54.843 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:54.843 "is_configured": true, 00:15:54.843 "data_offset": 256, 00:15:54.843 "data_size": 7936 00:15:54.843 } 00:15:54.843 ] 00:15:54.843 }' 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:54.843 [2024-12-05 19:06:12.282881] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:54.843 [2024-12-05 19:06:12.282993] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:54.843 [2024-12-05 19:06:12.283033] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:15:54.843 [2024-12-05 19:06:12.283043] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:54.843 [2024-12-05 19:06:12.283428] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:54.843 [2024-12-05 19:06:12.283448] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:54.843 [2024-12-05 19:06:12.283513] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:54.843 [2024-12-05 19:06:12.283530] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:54.843 [2024-12-05 19:06:12.283542] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:54.843 [2024-12-05 19:06:12.283553] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:54.843 BaseBdev1 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:54.843 19:06:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:55.783 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:55.783 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:55.783 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:55.783 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:55.783 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:55.783 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:55.783 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:55.783 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:55.783 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:55.783 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:55.783 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.783 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:55.783 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:55.783 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:55.783 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.043 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:56.043 "name": "raid_bdev1", 00:15:56.043 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:56.043 "strip_size_kb": 0, 00:15:56.043 "state": "online", 00:15:56.043 "raid_level": "raid1", 00:15:56.043 "superblock": true, 00:15:56.043 "num_base_bdevs": 2, 00:15:56.043 "num_base_bdevs_discovered": 1, 00:15:56.043 "num_base_bdevs_operational": 1, 00:15:56.043 "base_bdevs_list": [ 00:15:56.043 { 00:15:56.043 "name": null, 00:15:56.043 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:56.043 "is_configured": false, 00:15:56.043 "data_offset": 0, 00:15:56.043 "data_size": 7936 00:15:56.043 }, 00:15:56.043 { 00:15:56.043 "name": "BaseBdev2", 00:15:56.043 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:56.043 "is_configured": true, 00:15:56.043 "data_offset": 256, 00:15:56.043 "data_size": 7936 00:15:56.043 } 00:15:56.043 ] 00:15:56.043 }' 00:15:56.043 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:56.043 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:56.304 "name": "raid_bdev1", 00:15:56.304 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:56.304 "strip_size_kb": 0, 00:15:56.304 "state": "online", 00:15:56.304 "raid_level": "raid1", 00:15:56.304 "superblock": true, 00:15:56.304 "num_base_bdevs": 2, 00:15:56.304 "num_base_bdevs_discovered": 1, 00:15:56.304 "num_base_bdevs_operational": 1, 00:15:56.304 "base_bdevs_list": [ 00:15:56.304 { 00:15:56.304 "name": null, 00:15:56.304 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:56.304 "is_configured": false, 00:15:56.304 "data_offset": 0, 00:15:56.304 "data_size": 7936 00:15:56.304 }, 00:15:56.304 { 00:15:56.304 "name": "BaseBdev2", 00:15:56.304 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:56.304 "is_configured": true, 00:15:56.304 "data_offset": 256, 00:15:56.304 "data_size": 7936 00:15:56.304 } 00:15:56.304 ] 00:15:56.304 }' 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@652 -- # local es=0 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:56.304 [2024-12-05 19:06:13.828201] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:56.304 [2024-12-05 19:06:13.828353] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:56.304 [2024-12-05 19:06:13.828368] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:56.304 request: 00:15:56.304 { 00:15:56.304 "base_bdev": "BaseBdev1", 00:15:56.304 "raid_bdev": "raid_bdev1", 00:15:56.304 "method": "bdev_raid_add_base_bdev", 00:15:56.304 "req_id": 1 00:15:56.304 } 00:15:56.304 Got JSON-RPC error response 00:15:56.304 response: 00:15:56.304 { 00:15:56.304 "code": -22, 00:15:56.304 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:56.304 } 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@655 -- # es=1 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:15:56.304 19:06:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@779 -- # sleep 1 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:57.697 "name": "raid_bdev1", 00:15:57.697 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:57.697 "strip_size_kb": 0, 00:15:57.697 "state": "online", 00:15:57.697 "raid_level": "raid1", 00:15:57.697 "superblock": true, 00:15:57.697 "num_base_bdevs": 2, 00:15:57.697 "num_base_bdevs_discovered": 1, 00:15:57.697 "num_base_bdevs_operational": 1, 00:15:57.697 "base_bdevs_list": [ 00:15:57.697 { 00:15:57.697 "name": null, 00:15:57.697 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:57.697 "is_configured": false, 00:15:57.697 "data_offset": 0, 00:15:57.697 "data_size": 7936 00:15:57.697 }, 00:15:57.697 { 00:15:57.697 "name": "BaseBdev2", 00:15:57.697 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:57.697 "is_configured": true, 00:15:57.697 "data_offset": 256, 00:15:57.697 "data_size": 7936 00:15:57.697 } 00:15:57.697 ] 00:15:57.697 }' 00:15:57.697 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:57.698 19:06:14 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:57.957 "name": "raid_bdev1", 00:15:57.957 "uuid": "4c61b647-fc32-44e5-a223-8ff03dd5a450", 00:15:57.957 "strip_size_kb": 0, 00:15:57.957 "state": "online", 00:15:57.957 "raid_level": "raid1", 00:15:57.957 "superblock": true, 00:15:57.957 "num_base_bdevs": 2, 00:15:57.957 "num_base_bdevs_discovered": 1, 00:15:57.957 "num_base_bdevs_operational": 1, 00:15:57.957 "base_bdevs_list": [ 00:15:57.957 { 00:15:57.957 "name": null, 00:15:57.957 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:57.957 "is_configured": false, 00:15:57.957 "data_offset": 0, 00:15:57.957 "data_size": 7936 00:15:57.957 }, 00:15:57.957 { 00:15:57.957 "name": "BaseBdev2", 00:15:57.957 "uuid": "f0be5a7f-c080-527b-9061-ead324245372", 00:15:57.957 "is_configured": true, 00:15:57.957 "data_offset": 256, 00:15:57.957 "data_size": 7936 00:15:57.957 } 00:15:57.957 ] 00:15:57.957 }' 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@784 -- # killprocess 93996 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@954 -- # '[' -z 93996 ']' 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@958 -- # kill -0 93996 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@959 -- # uname 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 93996 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:57.957 killing process with pid 93996 00:15:57.957 Received shutdown signal, test time was about 60.000000 seconds 00:15:57.957 00:15:57.957 Latency(us) 00:15:57.957 [2024-12-05T19:06:15.516Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:57.957 [2024-12-05T19:06:15.516Z] =================================================================================================================== 00:15:57.957 [2024-12-05T19:06:15.516Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:57.957 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@972 -- # echo 'killing process with pid 93996' 00:15:57.958 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@973 -- # kill 93996 00:15:57.958 [2024-12-05 19:06:15.505893] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:57.958 [2024-12-05 19:06:15.505997] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:57.958 [2024-12-05 19:06:15.506045] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:57.958 [2024-12-05 19:06:15.506054] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:15:57.958 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@978 -- # wait 93996 00:15:58.217 [2024-12-05 19:06:15.537350] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:58.217 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@786 -- # return 0 00:15:58.217 00:15:58.217 real 0m18.402s 00:15:58.217 user 0m24.520s 00:15:58.217 sys 0m2.662s 00:15:58.217 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:58.217 19:06:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:58.217 ************************************ 00:15:58.217 END TEST raid_rebuild_test_sb_4k 00:15:58.217 ************************************ 00:15:58.475 19:06:15 bdev_raid -- bdev/bdev_raid.sh@1003 -- # base_malloc_params='-m 32' 00:15:58.476 19:06:15 bdev_raid -- bdev/bdev_raid.sh@1004 -- # run_test raid_state_function_test_sb_md_separate raid_state_function_test raid1 2 true 00:15:58.476 19:06:15 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:15:58.476 19:06:15 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:58.476 19:06:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:58.476 ************************************ 00:15:58.476 START TEST raid_state_function_test_sb_md_separate 00:15:58.476 ************************************ 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@229 -- # raid_pid=94568 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 94568' 00:15:58.476 Process raid pid: 94568 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@231 -- # waitforlisten 94568 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@835 -- # '[' -z 94568 ']' 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:58.476 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:58.476 19:06:15 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:58.476 [2024-12-05 19:06:15.917372] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:15:58.476 [2024-12-05 19:06:15.917571] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:58.735 [2024-12-05 19:06:16.074743] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:58.735 [2024-12-05 19:06:16.100750] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:58.735 [2024-12-05 19:06:16.144309] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:58.735 [2024-12-05 19:06:16.144421] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:59.304 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:59.304 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@868 -- # return 0 00:15:59.304 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:59.304 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.304 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:59.304 [2024-12-05 19:06:16.739357] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:59.304 [2024-12-05 19:06:16.739489] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:59.304 [2024-12-05 19:06:16.739521] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:59.304 [2024-12-05 19:06:16.739544] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:59.304 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.304 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:59.304 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:59.304 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:59.304 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:59.304 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:59.304 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:59.304 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:59.305 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:59.305 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:59.305 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:59.305 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:59.305 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.305 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:59.305 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:59.305 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.305 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:59.305 "name": "Existed_Raid", 00:15:59.305 "uuid": "f7c5ba81-0367-423d-9f68-ec3139382469", 00:15:59.305 "strip_size_kb": 0, 00:15:59.305 "state": "configuring", 00:15:59.305 "raid_level": "raid1", 00:15:59.305 "superblock": true, 00:15:59.305 "num_base_bdevs": 2, 00:15:59.305 "num_base_bdevs_discovered": 0, 00:15:59.305 "num_base_bdevs_operational": 2, 00:15:59.305 "base_bdevs_list": [ 00:15:59.305 { 00:15:59.305 "name": "BaseBdev1", 00:15:59.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:59.305 "is_configured": false, 00:15:59.305 "data_offset": 0, 00:15:59.305 "data_size": 0 00:15:59.305 }, 00:15:59.305 { 00:15:59.305 "name": "BaseBdev2", 00:15:59.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:59.305 "is_configured": false, 00:15:59.305 "data_offset": 0, 00:15:59.305 "data_size": 0 00:15:59.305 } 00:15:59.305 ] 00:15:59.305 }' 00:15:59.305 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:59.305 19:06:16 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:59.875 [2024-12-05 19:06:17.214461] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:59.875 [2024-12-05 19:06:17.214551] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:59.875 [2024-12-05 19:06:17.226449] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:59.875 [2024-12-05 19:06:17.226547] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:59.875 [2024-12-05 19:06:17.226572] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:59.875 [2024-12-05 19:06:17.226607] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:59.875 [2024-12-05 19:06:17.248020] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:59.875 BaseBdev1 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:15:59.875 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # local i 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:59.876 [ 00:15:59.876 { 00:15:59.876 "name": "BaseBdev1", 00:15:59.876 "aliases": [ 00:15:59.876 "8e0bf6eb-0745-4ca1-9529-8cd775a229b9" 00:15:59.876 ], 00:15:59.876 "product_name": "Malloc disk", 00:15:59.876 "block_size": 4096, 00:15:59.876 "num_blocks": 8192, 00:15:59.876 "uuid": "8e0bf6eb-0745-4ca1-9529-8cd775a229b9", 00:15:59.876 "md_size": 32, 00:15:59.876 "md_interleave": false, 00:15:59.876 "dif_type": 0, 00:15:59.876 "assigned_rate_limits": { 00:15:59.876 "rw_ios_per_sec": 0, 00:15:59.876 "rw_mbytes_per_sec": 0, 00:15:59.876 "r_mbytes_per_sec": 0, 00:15:59.876 "w_mbytes_per_sec": 0 00:15:59.876 }, 00:15:59.876 "claimed": true, 00:15:59.876 "claim_type": "exclusive_write", 00:15:59.876 "zoned": false, 00:15:59.876 "supported_io_types": { 00:15:59.876 "read": true, 00:15:59.876 "write": true, 00:15:59.876 "unmap": true, 00:15:59.876 "flush": true, 00:15:59.876 "reset": true, 00:15:59.876 "nvme_admin": false, 00:15:59.876 "nvme_io": false, 00:15:59.876 "nvme_io_md": false, 00:15:59.876 "write_zeroes": true, 00:15:59.876 "zcopy": true, 00:15:59.876 "get_zone_info": false, 00:15:59.876 "zone_management": false, 00:15:59.876 "zone_append": false, 00:15:59.876 "compare": false, 00:15:59.876 "compare_and_write": false, 00:15:59.876 "abort": true, 00:15:59.876 "seek_hole": false, 00:15:59.876 "seek_data": false, 00:15:59.876 "copy": true, 00:15:59.876 "nvme_iov_md": false 00:15:59.876 }, 00:15:59.876 "memory_domains": [ 00:15:59.876 { 00:15:59.876 "dma_device_id": "system", 00:15:59.876 "dma_device_type": 1 00:15:59.876 }, 00:15:59.876 { 00:15:59.876 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:59.876 "dma_device_type": 2 00:15:59.876 } 00:15:59.876 ], 00:15:59.876 "driver_specific": {} 00:15:59.876 } 00:15:59.876 ] 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@911 -- # return 0 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:59.876 "name": "Existed_Raid", 00:15:59.876 "uuid": "98612d3a-1705-4b65-92c0-6ceaf64a4ddf", 00:15:59.876 "strip_size_kb": 0, 00:15:59.876 "state": "configuring", 00:15:59.876 "raid_level": "raid1", 00:15:59.876 "superblock": true, 00:15:59.876 "num_base_bdevs": 2, 00:15:59.876 "num_base_bdevs_discovered": 1, 00:15:59.876 "num_base_bdevs_operational": 2, 00:15:59.876 "base_bdevs_list": [ 00:15:59.876 { 00:15:59.876 "name": "BaseBdev1", 00:15:59.876 "uuid": "8e0bf6eb-0745-4ca1-9529-8cd775a229b9", 00:15:59.876 "is_configured": true, 00:15:59.876 "data_offset": 256, 00:15:59.876 "data_size": 7936 00:15:59.876 }, 00:15:59.876 { 00:15:59.876 "name": "BaseBdev2", 00:15:59.876 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:59.876 "is_configured": false, 00:15:59.876 "data_offset": 0, 00:15:59.876 "data_size": 0 00:15:59.876 } 00:15:59.876 ] 00:15:59.876 }' 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:59.876 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.447 [2024-12-05 19:06:17.723249] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:00.447 [2024-12-05 19:06:17.723350] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.447 [2024-12-05 19:06:17.735285] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:00.447 [2024-12-05 19:06:17.737127] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:00.447 [2024-12-05 19:06:17.737173] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:00.447 "name": "Existed_Raid", 00:16:00.447 "uuid": "6f151c12-5333-409a-96e0-eede9b10fcad", 00:16:00.447 "strip_size_kb": 0, 00:16:00.447 "state": "configuring", 00:16:00.447 "raid_level": "raid1", 00:16:00.447 "superblock": true, 00:16:00.447 "num_base_bdevs": 2, 00:16:00.447 "num_base_bdevs_discovered": 1, 00:16:00.447 "num_base_bdevs_operational": 2, 00:16:00.447 "base_bdevs_list": [ 00:16:00.447 { 00:16:00.447 "name": "BaseBdev1", 00:16:00.447 "uuid": "8e0bf6eb-0745-4ca1-9529-8cd775a229b9", 00:16:00.447 "is_configured": true, 00:16:00.447 "data_offset": 256, 00:16:00.447 "data_size": 7936 00:16:00.447 }, 00:16:00.447 { 00:16:00.447 "name": "BaseBdev2", 00:16:00.447 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:00.447 "is_configured": false, 00:16:00.447 "data_offset": 0, 00:16:00.447 "data_size": 0 00:16:00.447 } 00:16:00.447 ] 00:16:00.447 }' 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:00.447 19:06:17 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.708 [2024-12-05 19:06:18.182105] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:00.708 [2024-12-05 19:06:18.182391] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:16:00.708 [2024-12-05 19:06:18.182447] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:00.708 [2024-12-05 19:06:18.182581] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:00.708 [2024-12-05 19:06:18.182740] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:16:00.708 [2024-12-05 19:06:18.182788] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:16:00.708 BaseBdev2 00:16:00.708 [2024-12-05 19:06:18.182897] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@905 -- # local i 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.708 [ 00:16:00.708 { 00:16:00.708 "name": "BaseBdev2", 00:16:00.708 "aliases": [ 00:16:00.708 "15e68c8e-5b8c-47c5-8cbc-aacdf79679c0" 00:16:00.708 ], 00:16:00.708 "product_name": "Malloc disk", 00:16:00.708 "block_size": 4096, 00:16:00.708 "num_blocks": 8192, 00:16:00.708 "uuid": "15e68c8e-5b8c-47c5-8cbc-aacdf79679c0", 00:16:00.708 "md_size": 32, 00:16:00.708 "md_interleave": false, 00:16:00.708 "dif_type": 0, 00:16:00.708 "assigned_rate_limits": { 00:16:00.708 "rw_ios_per_sec": 0, 00:16:00.708 "rw_mbytes_per_sec": 0, 00:16:00.708 "r_mbytes_per_sec": 0, 00:16:00.708 "w_mbytes_per_sec": 0 00:16:00.708 }, 00:16:00.708 "claimed": true, 00:16:00.708 "claim_type": "exclusive_write", 00:16:00.708 "zoned": false, 00:16:00.708 "supported_io_types": { 00:16:00.708 "read": true, 00:16:00.708 "write": true, 00:16:00.708 "unmap": true, 00:16:00.708 "flush": true, 00:16:00.708 "reset": true, 00:16:00.708 "nvme_admin": false, 00:16:00.708 "nvme_io": false, 00:16:00.708 "nvme_io_md": false, 00:16:00.708 "write_zeroes": true, 00:16:00.708 "zcopy": true, 00:16:00.708 "get_zone_info": false, 00:16:00.708 "zone_management": false, 00:16:00.708 "zone_append": false, 00:16:00.708 "compare": false, 00:16:00.708 "compare_and_write": false, 00:16:00.708 "abort": true, 00:16:00.708 "seek_hole": false, 00:16:00.708 "seek_data": false, 00:16:00.708 "copy": true, 00:16:00.708 "nvme_iov_md": false 00:16:00.708 }, 00:16:00.708 "memory_domains": [ 00:16:00.708 { 00:16:00.708 "dma_device_id": "system", 00:16:00.708 "dma_device_type": 1 00:16:00.708 }, 00:16:00.708 { 00:16:00.708 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:00.708 "dma_device_type": 2 00:16:00.708 } 00:16:00.708 ], 00:16:00.708 "driver_specific": {} 00:16:00.708 } 00:16:00.708 ] 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.708 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@911 -- # return 0 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.709 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:00.969 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:00.969 "name": "Existed_Raid", 00:16:00.969 "uuid": "6f151c12-5333-409a-96e0-eede9b10fcad", 00:16:00.969 "strip_size_kb": 0, 00:16:00.969 "state": "online", 00:16:00.969 "raid_level": "raid1", 00:16:00.969 "superblock": true, 00:16:00.969 "num_base_bdevs": 2, 00:16:00.969 "num_base_bdevs_discovered": 2, 00:16:00.969 "num_base_bdevs_operational": 2, 00:16:00.969 "base_bdevs_list": [ 00:16:00.969 { 00:16:00.969 "name": "BaseBdev1", 00:16:00.969 "uuid": "8e0bf6eb-0745-4ca1-9529-8cd775a229b9", 00:16:00.969 "is_configured": true, 00:16:00.969 "data_offset": 256, 00:16:00.969 "data_size": 7936 00:16:00.969 }, 00:16:00.969 { 00:16:00.969 "name": "BaseBdev2", 00:16:00.969 "uuid": "15e68c8e-5b8c-47c5-8cbc-aacdf79679c0", 00:16:00.969 "is_configured": true, 00:16:00.969 "data_offset": 256, 00:16:00.969 "data_size": 7936 00:16:00.969 } 00:16:00.969 ] 00:16:00.969 }' 00:16:00.970 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:00.970 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:01.231 [2024-12-05 19:06:18.645581] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:01.231 "name": "Existed_Raid", 00:16:01.231 "aliases": [ 00:16:01.231 "6f151c12-5333-409a-96e0-eede9b10fcad" 00:16:01.231 ], 00:16:01.231 "product_name": "Raid Volume", 00:16:01.231 "block_size": 4096, 00:16:01.231 "num_blocks": 7936, 00:16:01.231 "uuid": "6f151c12-5333-409a-96e0-eede9b10fcad", 00:16:01.231 "md_size": 32, 00:16:01.231 "md_interleave": false, 00:16:01.231 "dif_type": 0, 00:16:01.231 "assigned_rate_limits": { 00:16:01.231 "rw_ios_per_sec": 0, 00:16:01.231 "rw_mbytes_per_sec": 0, 00:16:01.231 "r_mbytes_per_sec": 0, 00:16:01.231 "w_mbytes_per_sec": 0 00:16:01.231 }, 00:16:01.231 "claimed": false, 00:16:01.231 "zoned": false, 00:16:01.231 "supported_io_types": { 00:16:01.231 "read": true, 00:16:01.231 "write": true, 00:16:01.231 "unmap": false, 00:16:01.231 "flush": false, 00:16:01.231 "reset": true, 00:16:01.231 "nvme_admin": false, 00:16:01.231 "nvme_io": false, 00:16:01.231 "nvme_io_md": false, 00:16:01.231 "write_zeroes": true, 00:16:01.231 "zcopy": false, 00:16:01.231 "get_zone_info": false, 00:16:01.231 "zone_management": false, 00:16:01.231 "zone_append": false, 00:16:01.231 "compare": false, 00:16:01.231 "compare_and_write": false, 00:16:01.231 "abort": false, 00:16:01.231 "seek_hole": false, 00:16:01.231 "seek_data": false, 00:16:01.231 "copy": false, 00:16:01.231 "nvme_iov_md": false 00:16:01.231 }, 00:16:01.231 "memory_domains": [ 00:16:01.231 { 00:16:01.231 "dma_device_id": "system", 00:16:01.231 "dma_device_type": 1 00:16:01.231 }, 00:16:01.231 { 00:16:01.231 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:01.231 "dma_device_type": 2 00:16:01.231 }, 00:16:01.231 { 00:16:01.231 "dma_device_id": "system", 00:16:01.231 "dma_device_type": 1 00:16:01.231 }, 00:16:01.231 { 00:16:01.231 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:01.231 "dma_device_type": 2 00:16:01.231 } 00:16:01.231 ], 00:16:01.231 "driver_specific": { 00:16:01.231 "raid": { 00:16:01.231 "uuid": "6f151c12-5333-409a-96e0-eede9b10fcad", 00:16:01.231 "strip_size_kb": 0, 00:16:01.231 "state": "online", 00:16:01.231 "raid_level": "raid1", 00:16:01.231 "superblock": true, 00:16:01.231 "num_base_bdevs": 2, 00:16:01.231 "num_base_bdevs_discovered": 2, 00:16:01.231 "num_base_bdevs_operational": 2, 00:16:01.231 "base_bdevs_list": [ 00:16:01.231 { 00:16:01.231 "name": "BaseBdev1", 00:16:01.231 "uuid": "8e0bf6eb-0745-4ca1-9529-8cd775a229b9", 00:16:01.231 "is_configured": true, 00:16:01.231 "data_offset": 256, 00:16:01.231 "data_size": 7936 00:16:01.231 }, 00:16:01.231 { 00:16:01.231 "name": "BaseBdev2", 00:16:01.231 "uuid": "15e68c8e-5b8c-47c5-8cbc-aacdf79679c0", 00:16:01.231 "is_configured": true, 00:16:01.231 "data_offset": 256, 00:16:01.231 "data_size": 7936 00:16:01.231 } 00:16:01.231 ] 00:16:01.231 } 00:16:01.231 } 00:16:01.231 }' 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:16:01.231 BaseBdev2' 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:16:01.231 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:01.492 [2024-12-05 19:06:18.876979] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@260 -- # local expected_state 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:01.492 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:01.493 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:01.493 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:01.493 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:01.493 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:01.493 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:01.493 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:01.493 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:01.493 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:01.493 "name": "Existed_Raid", 00:16:01.493 "uuid": "6f151c12-5333-409a-96e0-eede9b10fcad", 00:16:01.493 "strip_size_kb": 0, 00:16:01.493 "state": "online", 00:16:01.493 "raid_level": "raid1", 00:16:01.493 "superblock": true, 00:16:01.493 "num_base_bdevs": 2, 00:16:01.493 "num_base_bdevs_discovered": 1, 00:16:01.493 "num_base_bdevs_operational": 1, 00:16:01.493 "base_bdevs_list": [ 00:16:01.493 { 00:16:01.493 "name": null, 00:16:01.493 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:01.493 "is_configured": false, 00:16:01.493 "data_offset": 0, 00:16:01.493 "data_size": 7936 00:16:01.493 }, 00:16:01.493 { 00:16:01.493 "name": "BaseBdev2", 00:16:01.493 "uuid": "15e68c8e-5b8c-47c5-8cbc-aacdf79679c0", 00:16:01.493 "is_configured": true, 00:16:01.493 "data_offset": 256, 00:16:01.493 "data_size": 7936 00:16:01.493 } 00:16:01.493 ] 00:16:01.493 }' 00:16:01.493 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:01.493 19:06:18 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:01.752 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:16:01.752 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:01.752 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:01.752 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.012 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:02.012 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:16:02.012 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:02.013 [2024-12-05 19:06:19.364235] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:02.013 [2024-12-05 19:06:19.364387] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:02.013 [2024-12-05 19:06:19.376779] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:02.013 [2024-12-05 19:06:19.376902] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:02.013 [2024-12-05 19:06:19.376943] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@326 -- # killprocess 94568 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@954 -- # '[' -z 94568 ']' 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@958 -- # kill -0 94568 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@959 -- # uname 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 94568 00:16:02.013 killing process with pid 94568 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 94568' 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@973 -- # kill 94568 00:16:02.013 [2024-12-05 19:06:19.476962] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:02.013 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@978 -- # wait 94568 00:16:02.013 [2024-12-05 19:06:19.477988] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:02.274 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@328 -- # return 0 00:16:02.274 00:16:02.274 real 0m3.888s 00:16:02.274 user 0m6.112s 00:16:02.274 sys 0m0.839s 00:16:02.274 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:02.274 ************************************ 00:16:02.274 END TEST raid_state_function_test_sb_md_separate 00:16:02.274 ************************************ 00:16:02.274 19:06:19 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:02.274 19:06:19 bdev_raid -- bdev/bdev_raid.sh@1005 -- # run_test raid_superblock_test_md_separate raid_superblock_test raid1 2 00:16:02.274 19:06:19 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:16:02.274 19:06:19 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:02.274 19:06:19 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:02.274 ************************************ 00:16:02.274 START TEST raid_superblock_test_md_separate 00:16:02.274 ************************************ 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@399 -- # local strip_size 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@412 -- # raid_pid=94780 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@413 -- # waitforlisten 94780 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@835 -- # '[' -z 94780 ']' 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:02.274 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:02.274 19:06:19 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:02.535 [2024-12-05 19:06:19.881751] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:16:02.535 [2024-12-05 19:06:19.881952] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid94780 ] 00:16:02.535 [2024-12-05 19:06:20.029678] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:02.535 [2024-12-05 19:06:20.057005] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:02.796 [2024-12-05 19:06:20.101562] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:02.796 [2024-12-05 19:06:20.101688] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@868 -- # return 0 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc1 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:03.367 malloc1 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:03.367 [2024-12-05 19:06:20.730420] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:03.367 [2024-12-05 19:06:20.730508] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:03.367 [2024-12-05 19:06:20.730529] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:03.367 [2024-12-05 19:06:20.730540] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:03.367 [2024-12-05 19:06:20.732539] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:03.367 [2024-12-05 19:06:20.732586] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:03.367 pt1 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc2 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:03.367 malloc2 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.367 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:03.368 [2024-12-05 19:06:20.759772] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:03.368 [2024-12-05 19:06:20.759920] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:03.368 [2024-12-05 19:06:20.759952] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:03.368 [2024-12-05 19:06:20.759988] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:03.368 [2024-12-05 19:06:20.761911] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:03.368 [2024-12-05 19:06:20.762003] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:03.368 pt2 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:03.368 [2024-12-05 19:06:20.771784] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:03.368 [2024-12-05 19:06:20.773698] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:03.368 [2024-12-05 19:06:20.773902] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:03.368 [2024-12-05 19:06:20.773954] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:03.368 [2024-12-05 19:06:20.774064] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:03.368 [2024-12-05 19:06:20.774212] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:03.368 [2024-12-05 19:06:20.774253] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:03.368 [2024-12-05 19:06:20.774387] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:03.368 "name": "raid_bdev1", 00:16:03.368 "uuid": "534be6eb-4463-4b1f-9fd3-70fb90da593d", 00:16:03.368 "strip_size_kb": 0, 00:16:03.368 "state": "online", 00:16:03.368 "raid_level": "raid1", 00:16:03.368 "superblock": true, 00:16:03.368 "num_base_bdevs": 2, 00:16:03.368 "num_base_bdevs_discovered": 2, 00:16:03.368 "num_base_bdevs_operational": 2, 00:16:03.368 "base_bdevs_list": [ 00:16:03.368 { 00:16:03.368 "name": "pt1", 00:16:03.368 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:03.368 "is_configured": true, 00:16:03.368 "data_offset": 256, 00:16:03.368 "data_size": 7936 00:16:03.368 }, 00:16:03.368 { 00:16:03.368 "name": "pt2", 00:16:03.368 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:03.368 "is_configured": true, 00:16:03.368 "data_offset": 256, 00:16:03.368 "data_size": 7936 00:16:03.368 } 00:16:03.368 ] 00:16:03.368 }' 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:03.368 19:06:20 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:03.939 [2024-12-05 19:06:21.223268] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:03.939 "name": "raid_bdev1", 00:16:03.939 "aliases": [ 00:16:03.939 "534be6eb-4463-4b1f-9fd3-70fb90da593d" 00:16:03.939 ], 00:16:03.939 "product_name": "Raid Volume", 00:16:03.939 "block_size": 4096, 00:16:03.939 "num_blocks": 7936, 00:16:03.939 "uuid": "534be6eb-4463-4b1f-9fd3-70fb90da593d", 00:16:03.939 "md_size": 32, 00:16:03.939 "md_interleave": false, 00:16:03.939 "dif_type": 0, 00:16:03.939 "assigned_rate_limits": { 00:16:03.939 "rw_ios_per_sec": 0, 00:16:03.939 "rw_mbytes_per_sec": 0, 00:16:03.939 "r_mbytes_per_sec": 0, 00:16:03.939 "w_mbytes_per_sec": 0 00:16:03.939 }, 00:16:03.939 "claimed": false, 00:16:03.939 "zoned": false, 00:16:03.939 "supported_io_types": { 00:16:03.939 "read": true, 00:16:03.939 "write": true, 00:16:03.939 "unmap": false, 00:16:03.939 "flush": false, 00:16:03.939 "reset": true, 00:16:03.939 "nvme_admin": false, 00:16:03.939 "nvme_io": false, 00:16:03.939 "nvme_io_md": false, 00:16:03.939 "write_zeroes": true, 00:16:03.939 "zcopy": false, 00:16:03.939 "get_zone_info": false, 00:16:03.939 "zone_management": false, 00:16:03.939 "zone_append": false, 00:16:03.939 "compare": false, 00:16:03.939 "compare_and_write": false, 00:16:03.939 "abort": false, 00:16:03.939 "seek_hole": false, 00:16:03.939 "seek_data": false, 00:16:03.939 "copy": false, 00:16:03.939 "nvme_iov_md": false 00:16:03.939 }, 00:16:03.939 "memory_domains": [ 00:16:03.939 { 00:16:03.939 "dma_device_id": "system", 00:16:03.939 "dma_device_type": 1 00:16:03.939 }, 00:16:03.939 { 00:16:03.939 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:03.939 "dma_device_type": 2 00:16:03.939 }, 00:16:03.939 { 00:16:03.939 "dma_device_id": "system", 00:16:03.939 "dma_device_type": 1 00:16:03.939 }, 00:16:03.939 { 00:16:03.939 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:03.939 "dma_device_type": 2 00:16:03.939 } 00:16:03.939 ], 00:16:03.939 "driver_specific": { 00:16:03.939 "raid": { 00:16:03.939 "uuid": "534be6eb-4463-4b1f-9fd3-70fb90da593d", 00:16:03.939 "strip_size_kb": 0, 00:16:03.939 "state": "online", 00:16:03.939 "raid_level": "raid1", 00:16:03.939 "superblock": true, 00:16:03.939 "num_base_bdevs": 2, 00:16:03.939 "num_base_bdevs_discovered": 2, 00:16:03.939 "num_base_bdevs_operational": 2, 00:16:03.939 "base_bdevs_list": [ 00:16:03.939 { 00:16:03.939 "name": "pt1", 00:16:03.939 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:03.939 "is_configured": true, 00:16:03.939 "data_offset": 256, 00:16:03.939 "data_size": 7936 00:16:03.939 }, 00:16:03.939 { 00:16:03.939 "name": "pt2", 00:16:03.939 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:03.939 "is_configured": true, 00:16:03.939 "data_offset": 256, 00:16:03.939 "data_size": 7936 00:16:03.939 } 00:16:03.939 ] 00:16:03.939 } 00:16:03.939 } 00:16:03.939 }' 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:03.939 pt2' 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:03.939 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:03.940 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:03.940 [2024-12-05 19:06:21.478816] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=534be6eb-4463-4b1f-9fd3-70fb90da593d 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@436 -- # '[' -z 534be6eb-4463-4b1f-9fd3-70fb90da593d ']' 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.200 [2024-12-05 19:06:21.506544] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:04.200 [2024-12-05 19:06:21.506614] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:04.200 [2024-12-05 19:06:21.506742] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:04.200 [2024-12-05 19:06:21.506839] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:04.200 [2024-12-05 19:06:21.506887] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@652 -- # local es=0 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.200 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.200 [2024-12-05 19:06:21.650291] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:16:04.200 [2024-12-05 19:06:21.652118] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:16:04.200 [2024-12-05 19:06:21.652180] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:16:04.200 [2024-12-05 19:06:21.652222] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:16:04.200 [2024-12-05 19:06:21.652238] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:04.200 [2024-12-05 19:06:21.652246] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:16:04.200 request: 00:16:04.200 { 00:16:04.200 "name": "raid_bdev1", 00:16:04.200 "raid_level": "raid1", 00:16:04.200 "base_bdevs": [ 00:16:04.200 "malloc1", 00:16:04.200 "malloc2" 00:16:04.200 ], 00:16:04.200 "superblock": false, 00:16:04.200 "method": "bdev_raid_create", 00:16:04.200 "req_id": 1 00:16:04.200 } 00:16:04.200 Got JSON-RPC error response 00:16:04.200 response: 00:16:04.201 { 00:16:04.201 "code": -17, 00:16:04.201 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:16:04.201 } 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@655 -- # es=1 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.201 [2024-12-05 19:06:21.714158] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:04.201 [2024-12-05 19:06:21.714255] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:04.201 [2024-12-05 19:06:21.714304] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:04.201 [2024-12-05 19:06:21.714330] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:04.201 [2024-12-05 19:06:21.716221] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:04.201 [2024-12-05 19:06:21.716307] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:04.201 [2024-12-05 19:06:21.716394] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:04.201 [2024-12-05 19:06:21.716447] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:04.201 pt1 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.201 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.461 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:04.461 "name": "raid_bdev1", 00:16:04.461 "uuid": "534be6eb-4463-4b1f-9fd3-70fb90da593d", 00:16:04.461 "strip_size_kb": 0, 00:16:04.461 "state": "configuring", 00:16:04.461 "raid_level": "raid1", 00:16:04.461 "superblock": true, 00:16:04.461 "num_base_bdevs": 2, 00:16:04.461 "num_base_bdevs_discovered": 1, 00:16:04.461 "num_base_bdevs_operational": 2, 00:16:04.461 "base_bdevs_list": [ 00:16:04.461 { 00:16:04.461 "name": "pt1", 00:16:04.461 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:04.461 "is_configured": true, 00:16:04.461 "data_offset": 256, 00:16:04.461 "data_size": 7936 00:16:04.461 }, 00:16:04.461 { 00:16:04.461 "name": null, 00:16:04.461 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:04.461 "is_configured": false, 00:16:04.461 "data_offset": 256, 00:16:04.461 "data_size": 7936 00:16:04.461 } 00:16:04.461 ] 00:16:04.461 }' 00:16:04.461 19:06:21 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:04.461 19:06:21 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.723 [2024-12-05 19:06:22.145421] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:04.723 [2024-12-05 19:06:22.145514] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:04.723 [2024-12-05 19:06:22.145534] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:16:04.723 [2024-12-05 19:06:22.145542] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:04.723 [2024-12-05 19:06:22.145683] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:04.723 [2024-12-05 19:06:22.145698] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:04.723 [2024-12-05 19:06:22.145738] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:04.723 [2024-12-05 19:06:22.145760] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:04.723 [2024-12-05 19:06:22.145848] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:16:04.723 [2024-12-05 19:06:22.145856] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:04.723 [2024-12-05 19:06:22.145921] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:16:04.723 [2024-12-05 19:06:22.145997] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:16:04.723 [2024-12-05 19:06:22.146009] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:16:04.723 [2024-12-05 19:06:22.146065] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:04.723 pt2 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:04.723 "name": "raid_bdev1", 00:16:04.723 "uuid": "534be6eb-4463-4b1f-9fd3-70fb90da593d", 00:16:04.723 "strip_size_kb": 0, 00:16:04.723 "state": "online", 00:16:04.723 "raid_level": "raid1", 00:16:04.723 "superblock": true, 00:16:04.723 "num_base_bdevs": 2, 00:16:04.723 "num_base_bdevs_discovered": 2, 00:16:04.723 "num_base_bdevs_operational": 2, 00:16:04.723 "base_bdevs_list": [ 00:16:04.723 { 00:16:04.723 "name": "pt1", 00:16:04.723 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:04.723 "is_configured": true, 00:16:04.723 "data_offset": 256, 00:16:04.723 "data_size": 7936 00:16:04.723 }, 00:16:04.723 { 00:16:04.723 "name": "pt2", 00:16:04.723 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:04.723 "is_configured": true, 00:16:04.723 "data_offset": 256, 00:16:04.723 "data_size": 7936 00:16:04.723 } 00:16:04.723 ] 00:16:04.723 }' 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:04.723 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.294 [2024-12-05 19:06:22.612890] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:05.294 "name": "raid_bdev1", 00:16:05.294 "aliases": [ 00:16:05.294 "534be6eb-4463-4b1f-9fd3-70fb90da593d" 00:16:05.294 ], 00:16:05.294 "product_name": "Raid Volume", 00:16:05.294 "block_size": 4096, 00:16:05.294 "num_blocks": 7936, 00:16:05.294 "uuid": "534be6eb-4463-4b1f-9fd3-70fb90da593d", 00:16:05.294 "md_size": 32, 00:16:05.294 "md_interleave": false, 00:16:05.294 "dif_type": 0, 00:16:05.294 "assigned_rate_limits": { 00:16:05.294 "rw_ios_per_sec": 0, 00:16:05.294 "rw_mbytes_per_sec": 0, 00:16:05.294 "r_mbytes_per_sec": 0, 00:16:05.294 "w_mbytes_per_sec": 0 00:16:05.294 }, 00:16:05.294 "claimed": false, 00:16:05.294 "zoned": false, 00:16:05.294 "supported_io_types": { 00:16:05.294 "read": true, 00:16:05.294 "write": true, 00:16:05.294 "unmap": false, 00:16:05.294 "flush": false, 00:16:05.294 "reset": true, 00:16:05.294 "nvme_admin": false, 00:16:05.294 "nvme_io": false, 00:16:05.294 "nvme_io_md": false, 00:16:05.294 "write_zeroes": true, 00:16:05.294 "zcopy": false, 00:16:05.294 "get_zone_info": false, 00:16:05.294 "zone_management": false, 00:16:05.294 "zone_append": false, 00:16:05.294 "compare": false, 00:16:05.294 "compare_and_write": false, 00:16:05.294 "abort": false, 00:16:05.294 "seek_hole": false, 00:16:05.294 "seek_data": false, 00:16:05.294 "copy": false, 00:16:05.294 "nvme_iov_md": false 00:16:05.294 }, 00:16:05.294 "memory_domains": [ 00:16:05.294 { 00:16:05.294 "dma_device_id": "system", 00:16:05.294 "dma_device_type": 1 00:16:05.294 }, 00:16:05.294 { 00:16:05.294 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:05.294 "dma_device_type": 2 00:16:05.294 }, 00:16:05.294 { 00:16:05.294 "dma_device_id": "system", 00:16:05.294 "dma_device_type": 1 00:16:05.294 }, 00:16:05.294 { 00:16:05.294 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:05.294 "dma_device_type": 2 00:16:05.294 } 00:16:05.294 ], 00:16:05.294 "driver_specific": { 00:16:05.294 "raid": { 00:16:05.294 "uuid": "534be6eb-4463-4b1f-9fd3-70fb90da593d", 00:16:05.294 "strip_size_kb": 0, 00:16:05.294 "state": "online", 00:16:05.294 "raid_level": "raid1", 00:16:05.294 "superblock": true, 00:16:05.294 "num_base_bdevs": 2, 00:16:05.294 "num_base_bdevs_discovered": 2, 00:16:05.294 "num_base_bdevs_operational": 2, 00:16:05.294 "base_bdevs_list": [ 00:16:05.294 { 00:16:05.294 "name": "pt1", 00:16:05.294 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:05.294 "is_configured": true, 00:16:05.294 "data_offset": 256, 00:16:05.294 "data_size": 7936 00:16:05.294 }, 00:16:05.294 { 00:16:05.294 "name": "pt2", 00:16:05.294 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:05.294 "is_configured": true, 00:16:05.294 "data_offset": 256, 00:16:05.294 "data_size": 7936 00:16:05.294 } 00:16:05.294 ] 00:16:05.294 } 00:16:05.294 } 00:16:05.294 }' 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:05.294 pt2' 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.294 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.295 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.295 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:05.295 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:05.295 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:05.295 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:05.295 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.295 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.295 [2024-12-05 19:06:22.832459] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # '[' 534be6eb-4463-4b1f-9fd3-70fb90da593d '!=' 534be6eb-4463-4b1f-9fd3-70fb90da593d ']' 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.555 [2024-12-05 19:06:22.876180] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:05.555 "name": "raid_bdev1", 00:16:05.555 "uuid": "534be6eb-4463-4b1f-9fd3-70fb90da593d", 00:16:05.555 "strip_size_kb": 0, 00:16:05.555 "state": "online", 00:16:05.555 "raid_level": "raid1", 00:16:05.555 "superblock": true, 00:16:05.555 "num_base_bdevs": 2, 00:16:05.555 "num_base_bdevs_discovered": 1, 00:16:05.555 "num_base_bdevs_operational": 1, 00:16:05.555 "base_bdevs_list": [ 00:16:05.555 { 00:16:05.555 "name": null, 00:16:05.555 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:05.555 "is_configured": false, 00:16:05.555 "data_offset": 0, 00:16:05.555 "data_size": 7936 00:16:05.555 }, 00:16:05.555 { 00:16:05.555 "name": "pt2", 00:16:05.555 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:05.555 "is_configured": true, 00:16:05.555 "data_offset": 256, 00:16:05.555 "data_size": 7936 00:16:05.555 } 00:16:05.555 ] 00:16:05.555 }' 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:05.555 19:06:22 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.816 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:05.816 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.816 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.816 [2024-12-05 19:06:23.327399] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:05.816 [2024-12-05 19:06:23.327470] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:05.816 [2024-12-05 19:06:23.327554] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:05.816 [2024-12-05 19:06:23.327609] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:05.816 [2024-12-05 19:06:23.327656] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:16:05.816 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:05.816 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:05.816 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:05.816 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:05.816 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.816 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@519 -- # i=1 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.076 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.076 [2024-12-05 19:06:23.403273] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:06.076 [2024-12-05 19:06:23.403319] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:06.076 [2024-12-05 19:06:23.403344] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:16:06.076 [2024-12-05 19:06:23.403353] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:06.076 [2024-12-05 19:06:23.405280] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:06.076 [2024-12-05 19:06:23.405355] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:06.076 [2024-12-05 19:06:23.405430] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:06.076 [2024-12-05 19:06:23.405462] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:06.076 [2024-12-05 19:06:23.405533] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:16:06.076 [2024-12-05 19:06:23.405540] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:06.076 [2024-12-05 19:06:23.405604] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:16:06.076 [2024-12-05 19:06:23.405676] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:16:06.077 [2024-12-05 19:06:23.405699] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:16:06.077 [2024-12-05 19:06:23.405761] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:06.077 pt2 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:06.077 "name": "raid_bdev1", 00:16:06.077 "uuid": "534be6eb-4463-4b1f-9fd3-70fb90da593d", 00:16:06.077 "strip_size_kb": 0, 00:16:06.077 "state": "online", 00:16:06.077 "raid_level": "raid1", 00:16:06.077 "superblock": true, 00:16:06.077 "num_base_bdevs": 2, 00:16:06.077 "num_base_bdevs_discovered": 1, 00:16:06.077 "num_base_bdevs_operational": 1, 00:16:06.077 "base_bdevs_list": [ 00:16:06.077 { 00:16:06.077 "name": null, 00:16:06.077 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:06.077 "is_configured": false, 00:16:06.077 "data_offset": 256, 00:16:06.077 "data_size": 7936 00:16:06.077 }, 00:16:06.077 { 00:16:06.077 "name": "pt2", 00:16:06.077 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:06.077 "is_configured": true, 00:16:06.077 "data_offset": 256, 00:16:06.077 "data_size": 7936 00:16:06.077 } 00:16:06.077 ] 00:16:06.077 }' 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:06.077 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.337 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:06.337 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.337 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.337 [2024-12-05 19:06:23.858471] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:06.337 [2024-12-05 19:06:23.858554] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:06.337 [2024-12-05 19:06:23.858634] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:06.337 [2024-12-05 19:06:23.858703] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:06.337 [2024-12-05 19:06:23.858758] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:16:06.337 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.337 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.337 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:06.337 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.337 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.337 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.603 [2024-12-05 19:06:23.918385] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:06.603 [2024-12-05 19:06:23.918516] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:06.603 [2024-12-05 19:06:23.918537] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:16:06.603 [2024-12-05 19:06:23.918551] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:06.603 [2024-12-05 19:06:23.920486] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:06.603 [2024-12-05 19:06:23.920523] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:06.603 [2024-12-05 19:06:23.920583] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:06.603 [2024-12-05 19:06:23.920618] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:06.603 [2024-12-05 19:06:23.920723] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:06.603 [2024-12-05 19:06:23.920738] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:06.603 [2024-12-05 19:06:23.920751] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:16:06.603 [2024-12-05 19:06:23.920790] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:06.603 [2024-12-05 19:06:23.920847] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:16:06.603 [2024-12-05 19:06:23.920868] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:06.603 [2024-12-05 19:06:23.920936] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:06.603 [2024-12-05 19:06:23.921010] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:16:06.603 [2024-12-05 19:06:23.921026] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:16:06.603 [2024-12-05 19:06:23.921099] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:06.603 pt1 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:06.603 "name": "raid_bdev1", 00:16:06.603 "uuid": "534be6eb-4463-4b1f-9fd3-70fb90da593d", 00:16:06.603 "strip_size_kb": 0, 00:16:06.603 "state": "online", 00:16:06.603 "raid_level": "raid1", 00:16:06.603 "superblock": true, 00:16:06.603 "num_base_bdevs": 2, 00:16:06.603 "num_base_bdevs_discovered": 1, 00:16:06.603 "num_base_bdevs_operational": 1, 00:16:06.603 "base_bdevs_list": [ 00:16:06.603 { 00:16:06.603 "name": null, 00:16:06.603 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:06.603 "is_configured": false, 00:16:06.603 "data_offset": 256, 00:16:06.603 "data_size": 7936 00:16:06.603 }, 00:16:06.603 { 00:16:06.603 "name": "pt2", 00:16:06.603 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:06.603 "is_configured": true, 00:16:06.603 "data_offset": 256, 00:16:06.603 "data_size": 7936 00:16:06.603 } 00:16:06.603 ] 00:16:06.603 }' 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:06.603 19:06:23 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.881 19:06:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:16:06.881 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:06.881 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.881 19:06:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:06.881 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.177 19:06:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:16:07.177 19:06:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:07.177 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:07.177 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:07.177 19:06:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:16:07.177 [2024-12-05 19:06:24.461707] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:07.177 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:07.178 19:06:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # '[' 534be6eb-4463-4b1f-9fd3-70fb90da593d '!=' 534be6eb-4463-4b1f-9fd3-70fb90da593d ']' 00:16:07.178 19:06:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@563 -- # killprocess 94780 00:16:07.178 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@954 -- # '[' -z 94780 ']' 00:16:07.178 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@958 -- # kill -0 94780 00:16:07.178 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@959 -- # uname 00:16:07.178 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:07.178 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 94780 00:16:07.178 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:07.178 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:07.178 killing process with pid 94780 00:16:07.178 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 94780' 00:16:07.178 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@973 -- # kill 94780 00:16:07.178 [2024-12-05 19:06:24.546091] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:07.178 [2024-12-05 19:06:24.546159] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:07.178 [2024-12-05 19:06:24.546201] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:07.178 [2024-12-05 19:06:24.546209] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:16:07.178 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@978 -- # wait 94780 00:16:07.178 [2024-12-05 19:06:24.570693] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:07.441 ************************************ 00:16:07.441 END TEST raid_superblock_test_md_separate 00:16:07.441 ************************************ 00:16:07.441 19:06:24 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@565 -- # return 0 00:16:07.441 00:16:07.441 real 0m4.991s 00:16:07.441 user 0m8.157s 00:16:07.441 sys 0m1.120s 00:16:07.441 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:07.441 19:06:24 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:07.441 19:06:24 bdev_raid -- bdev/bdev_raid.sh@1006 -- # '[' true = true ']' 00:16:07.441 19:06:24 bdev_raid -- bdev/bdev_raid.sh@1007 -- # run_test raid_rebuild_test_sb_md_separate raid_rebuild_test raid1 2 true false true 00:16:07.441 19:06:24 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:16:07.441 19:06:24 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:07.441 19:06:24 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:07.441 ************************************ 00:16:07.441 START TEST raid_rebuild_test_sb_md_separate 00:16:07.441 ************************************ 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false true 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # local verify=true 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@597 -- # raid_pid=95067 00:16:07.441 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@598 -- # waitforlisten 95067 00:16:07.442 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:07.442 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@835 -- # '[' -z 95067 ']' 00:16:07.442 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:07.442 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:07.442 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:07.442 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:07.442 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:07.442 19:06:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:07.442 I/O size of 3145728 is greater than zero copy threshold (65536). 00:16:07.442 Zero copy mechanism will not be used. 00:16:07.442 [2024-12-05 19:06:24.959962] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:16:07.442 [2024-12-05 19:06:24.960103] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid95067 ] 00:16:07.700 [2024-12-05 19:06:25.117849] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:07.700 [2024-12-05 19:06:25.143700] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:07.700 [2024-12-05 19:06:25.187142] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:07.700 [2024-12-05 19:06:25.187175] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # return 0 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1_malloc 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.267 BaseBdev1_malloc 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.267 [2024-12-05 19:06:25.791411] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:08.267 [2024-12-05 19:06:25.791479] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:08.267 [2024-12-05 19:06:25.791522] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:08.267 [2024-12-05 19:06:25.791540] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:08.267 [2024-12-05 19:06:25.793520] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:08.267 [2024-12-05 19:06:25.793556] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:08.267 BaseBdev1 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2_malloc 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.267 BaseBdev2_malloc 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.267 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.267 [2024-12-05 19:06:25.820713] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:08.267 [2024-12-05 19:06:25.820773] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:08.267 [2024-12-05 19:06:25.820795] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:08.267 [2024-12-05 19:06:25.820804] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:08.267 [2024-12-05 19:06:25.822686] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:08.267 [2024-12-05 19:06:25.822719] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:08.526 BaseBdev2 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b spare_malloc 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.526 spare_malloc 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.526 spare_delay 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.526 [2024-12-05 19:06:25.875384] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:08.526 [2024-12-05 19:06:25.875458] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:08.526 [2024-12-05 19:06:25.875492] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:16:08.526 [2024-12-05 19:06:25.875506] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:08.526 [2024-12-05 19:06:25.878477] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:08.526 [2024-12-05 19:06:25.878529] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:08.526 spare 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.526 [2024-12-05 19:06:25.887478] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:08.526 [2024-12-05 19:06:25.889542] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:08.526 [2024-12-05 19:06:25.889811] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:08.526 [2024-12-05 19:06:25.889831] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:08.526 [2024-12-05 19:06:25.889924] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:16:08.526 [2024-12-05 19:06:25.890047] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:08.526 [2024-12-05 19:06:25.890060] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:08.526 [2024-12-05 19:06:25.890139] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:08.526 "name": "raid_bdev1", 00:16:08.526 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:08.526 "strip_size_kb": 0, 00:16:08.526 "state": "online", 00:16:08.526 "raid_level": "raid1", 00:16:08.526 "superblock": true, 00:16:08.526 "num_base_bdevs": 2, 00:16:08.526 "num_base_bdevs_discovered": 2, 00:16:08.526 "num_base_bdevs_operational": 2, 00:16:08.526 "base_bdevs_list": [ 00:16:08.526 { 00:16:08.526 "name": "BaseBdev1", 00:16:08.526 "uuid": "2d51f346-9aa7-5f17-9130-ce0b2fb6b505", 00:16:08.526 "is_configured": true, 00:16:08.526 "data_offset": 256, 00:16:08.526 "data_size": 7936 00:16:08.526 }, 00:16:08.526 { 00:16:08.526 "name": "BaseBdev2", 00:16:08.526 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:08.526 "is_configured": true, 00:16:08.526 "data_offset": 256, 00:16:08.526 "data_size": 7936 00:16:08.526 } 00:16:08.526 ] 00:16:08.526 }' 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:08.526 19:06:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.785 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:08.785 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:08.785 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:08.785 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.785 [2024-12-05 19:06:26.338881] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:09.044 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:16:09.303 [2024-12-05 19:06:26.610246] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:09.303 /dev/nbd0 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:09.303 1+0 records in 00:16:09.303 1+0 records out 00:16:09.303 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000406567 s, 10.1 MB/s 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:09.303 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:16:09.304 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:09.304 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:09.304 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:16:09.304 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:16:09.304 19:06:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:16:09.873 7936+0 records in 00:16:09.873 7936+0 records out 00:16:09.873 32505856 bytes (33 MB, 31 MiB) copied, 0.556201 s, 58.4 MB/s 00:16:09.873 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:16:09.873 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:09.873 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:09.873 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:09.873 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:16:09.873 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:09.873 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:10.132 [2024-12-05 19:06:27.461732] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:10.132 [2024-12-05 19:06:27.477832] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:10.132 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:10.133 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:10.133 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:10.133 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:10.133 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:10.133 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:10.133 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:10.133 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:10.133 "name": "raid_bdev1", 00:16:10.133 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:10.133 "strip_size_kb": 0, 00:16:10.133 "state": "online", 00:16:10.133 "raid_level": "raid1", 00:16:10.133 "superblock": true, 00:16:10.133 "num_base_bdevs": 2, 00:16:10.133 "num_base_bdevs_discovered": 1, 00:16:10.133 "num_base_bdevs_operational": 1, 00:16:10.133 "base_bdevs_list": [ 00:16:10.133 { 00:16:10.133 "name": null, 00:16:10.133 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:10.133 "is_configured": false, 00:16:10.133 "data_offset": 0, 00:16:10.133 "data_size": 7936 00:16:10.133 }, 00:16:10.133 { 00:16:10.133 "name": "BaseBdev2", 00:16:10.133 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:10.133 "is_configured": true, 00:16:10.133 "data_offset": 256, 00:16:10.133 "data_size": 7936 00:16:10.133 } 00:16:10.133 ] 00:16:10.133 }' 00:16:10.133 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:10.133 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:10.701 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:10.701 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:10.701 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:10.701 [2024-12-05 19:06:27.956994] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:10.701 [2024-12-05 19:06:27.961621] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019c960 00:16:10.701 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:10.701 19:06:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:10.701 [2024-12-05 19:06:27.963889] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:11.641 19:06:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:11.641 19:06:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:11.641 19:06:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:11.641 19:06:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:11.641 19:06:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:11.641 19:06:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:11.641 19:06:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:11.641 19:06:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.641 19:06:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:11.641 19:06:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.641 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:11.641 "name": "raid_bdev1", 00:16:11.641 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:11.641 "strip_size_kb": 0, 00:16:11.641 "state": "online", 00:16:11.641 "raid_level": "raid1", 00:16:11.641 "superblock": true, 00:16:11.641 "num_base_bdevs": 2, 00:16:11.641 "num_base_bdevs_discovered": 2, 00:16:11.641 "num_base_bdevs_operational": 2, 00:16:11.641 "process": { 00:16:11.641 "type": "rebuild", 00:16:11.641 "target": "spare", 00:16:11.641 "progress": { 00:16:11.641 "blocks": 2560, 00:16:11.641 "percent": 32 00:16:11.641 } 00:16:11.641 }, 00:16:11.641 "base_bdevs_list": [ 00:16:11.641 { 00:16:11.641 "name": "spare", 00:16:11.641 "uuid": "284dae40-c263-5a17-a73f-a8e0353939e1", 00:16:11.641 "is_configured": true, 00:16:11.641 "data_offset": 256, 00:16:11.641 "data_size": 7936 00:16:11.641 }, 00:16:11.641 { 00:16:11.641 "name": "BaseBdev2", 00:16:11.641 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:11.641 "is_configured": true, 00:16:11.641 "data_offset": 256, 00:16:11.641 "data_size": 7936 00:16:11.641 } 00:16:11.641 ] 00:16:11.641 }' 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:11.642 [2024-12-05 19:06:29.120758] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:11.642 [2024-12-05 19:06:29.172635] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:11.642 [2024-12-05 19:06:29.172759] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:11.642 [2024-12-05 19:06:29.172784] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:11.642 [2024-12-05 19:06:29.172793] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:11.642 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:11.900 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:11.900 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:11.901 "name": "raid_bdev1", 00:16:11.901 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:11.901 "strip_size_kb": 0, 00:16:11.901 "state": "online", 00:16:11.901 "raid_level": "raid1", 00:16:11.901 "superblock": true, 00:16:11.901 "num_base_bdevs": 2, 00:16:11.901 "num_base_bdevs_discovered": 1, 00:16:11.901 "num_base_bdevs_operational": 1, 00:16:11.901 "base_bdevs_list": [ 00:16:11.901 { 00:16:11.901 "name": null, 00:16:11.901 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:11.901 "is_configured": false, 00:16:11.901 "data_offset": 0, 00:16:11.901 "data_size": 7936 00:16:11.901 }, 00:16:11.901 { 00:16:11.901 "name": "BaseBdev2", 00:16:11.901 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:11.901 "is_configured": true, 00:16:11.901 "data_offset": 256, 00:16:11.901 "data_size": 7936 00:16:11.901 } 00:16:11.901 ] 00:16:11.901 }' 00:16:11.901 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:11.901 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.159 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:12.159 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:12.159 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:12.159 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:12.159 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:12.159 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:12.159 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.159 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:12.159 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.159 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.159 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:12.159 "name": "raid_bdev1", 00:16:12.159 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:12.159 "strip_size_kb": 0, 00:16:12.159 "state": "online", 00:16:12.159 "raid_level": "raid1", 00:16:12.159 "superblock": true, 00:16:12.159 "num_base_bdevs": 2, 00:16:12.159 "num_base_bdevs_discovered": 1, 00:16:12.160 "num_base_bdevs_operational": 1, 00:16:12.160 "base_bdevs_list": [ 00:16:12.160 { 00:16:12.160 "name": null, 00:16:12.160 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:12.160 "is_configured": false, 00:16:12.160 "data_offset": 0, 00:16:12.160 "data_size": 7936 00:16:12.160 }, 00:16:12.160 { 00:16:12.160 "name": "BaseBdev2", 00:16:12.160 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:12.160 "is_configured": true, 00:16:12.160 "data_offset": 256, 00:16:12.160 "data_size": 7936 00:16:12.160 } 00:16:12.160 ] 00:16:12.160 }' 00:16:12.160 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:12.418 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:12.418 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:12.418 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:12.418 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:12.418 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:12.418 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.418 [2024-12-05 19:06:29.808676] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:12.418 [2024-12-05 19:06:29.812314] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019ca30 00:16:12.418 [2024-12-05 19:06:29.814469] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:12.418 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:12.418 19:06:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:13.359 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:13.359 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:13.359 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:13.359 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:13.360 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:13.360 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:13.360 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:13.360 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:13.360 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:13.360 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:13.360 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:13.360 "name": "raid_bdev1", 00:16:13.360 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:13.360 "strip_size_kb": 0, 00:16:13.360 "state": "online", 00:16:13.360 "raid_level": "raid1", 00:16:13.360 "superblock": true, 00:16:13.360 "num_base_bdevs": 2, 00:16:13.360 "num_base_bdevs_discovered": 2, 00:16:13.360 "num_base_bdevs_operational": 2, 00:16:13.360 "process": { 00:16:13.360 "type": "rebuild", 00:16:13.360 "target": "spare", 00:16:13.360 "progress": { 00:16:13.360 "blocks": 2560, 00:16:13.360 "percent": 32 00:16:13.360 } 00:16:13.360 }, 00:16:13.360 "base_bdevs_list": [ 00:16:13.360 { 00:16:13.360 "name": "spare", 00:16:13.360 "uuid": "284dae40-c263-5a17-a73f-a8e0353939e1", 00:16:13.360 "is_configured": true, 00:16:13.360 "data_offset": 256, 00:16:13.360 "data_size": 7936 00:16:13.360 }, 00:16:13.360 { 00:16:13.360 "name": "BaseBdev2", 00:16:13.360 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:13.360 "is_configured": true, 00:16:13.360 "data_offset": 256, 00:16:13.360 "data_size": 7936 00:16:13.360 } 00:16:13.360 ] 00:16:13.360 }' 00:16:13.360 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:16:13.619 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@706 -- # local timeout=580 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:13.619 "name": "raid_bdev1", 00:16:13.619 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:13.619 "strip_size_kb": 0, 00:16:13.619 "state": "online", 00:16:13.619 "raid_level": "raid1", 00:16:13.619 "superblock": true, 00:16:13.619 "num_base_bdevs": 2, 00:16:13.619 "num_base_bdevs_discovered": 2, 00:16:13.619 "num_base_bdevs_operational": 2, 00:16:13.619 "process": { 00:16:13.619 "type": "rebuild", 00:16:13.619 "target": "spare", 00:16:13.619 "progress": { 00:16:13.619 "blocks": 2816, 00:16:13.619 "percent": 35 00:16:13.619 } 00:16:13.619 }, 00:16:13.619 "base_bdevs_list": [ 00:16:13.619 { 00:16:13.619 "name": "spare", 00:16:13.619 "uuid": "284dae40-c263-5a17-a73f-a8e0353939e1", 00:16:13.619 "is_configured": true, 00:16:13.619 "data_offset": 256, 00:16:13.619 "data_size": 7936 00:16:13.619 }, 00:16:13.619 { 00:16:13.619 "name": "BaseBdev2", 00:16:13.619 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:13.619 "is_configured": true, 00:16:13.619 "data_offset": 256, 00:16:13.619 "data_size": 7936 00:16:13.619 } 00:16:13.619 ] 00:16:13.619 }' 00:16:13.619 19:06:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:13.619 19:06:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:13.619 19:06:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:13.619 19:06:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:13.619 19:06:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:14.555 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:14.555 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:14.555 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:14.555 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:14.555 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:14.555 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:14.555 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:14.555 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:14.555 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:14.555 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:14.815 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:14.815 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:14.815 "name": "raid_bdev1", 00:16:14.815 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:14.815 "strip_size_kb": 0, 00:16:14.815 "state": "online", 00:16:14.815 "raid_level": "raid1", 00:16:14.815 "superblock": true, 00:16:14.815 "num_base_bdevs": 2, 00:16:14.815 "num_base_bdevs_discovered": 2, 00:16:14.815 "num_base_bdevs_operational": 2, 00:16:14.815 "process": { 00:16:14.815 "type": "rebuild", 00:16:14.815 "target": "spare", 00:16:14.815 "progress": { 00:16:14.815 "blocks": 5632, 00:16:14.815 "percent": 70 00:16:14.815 } 00:16:14.815 }, 00:16:14.815 "base_bdevs_list": [ 00:16:14.815 { 00:16:14.815 "name": "spare", 00:16:14.815 "uuid": "284dae40-c263-5a17-a73f-a8e0353939e1", 00:16:14.815 "is_configured": true, 00:16:14.815 "data_offset": 256, 00:16:14.815 "data_size": 7936 00:16:14.815 }, 00:16:14.815 { 00:16:14.815 "name": "BaseBdev2", 00:16:14.815 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:14.815 "is_configured": true, 00:16:14.815 "data_offset": 256, 00:16:14.815 "data_size": 7936 00:16:14.815 } 00:16:14.815 ] 00:16:14.815 }' 00:16:14.815 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:14.815 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:14.815 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:14.815 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:14.815 19:06:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:15.384 [2024-12-05 19:06:32.929273] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:15.384 [2024-12-05 19:06:32.929441] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:15.384 [2024-12-05 19:06:32.929592] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:15.954 "name": "raid_bdev1", 00:16:15.954 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:15.954 "strip_size_kb": 0, 00:16:15.954 "state": "online", 00:16:15.954 "raid_level": "raid1", 00:16:15.954 "superblock": true, 00:16:15.954 "num_base_bdevs": 2, 00:16:15.954 "num_base_bdevs_discovered": 2, 00:16:15.954 "num_base_bdevs_operational": 2, 00:16:15.954 "base_bdevs_list": [ 00:16:15.954 { 00:16:15.954 "name": "spare", 00:16:15.954 "uuid": "284dae40-c263-5a17-a73f-a8e0353939e1", 00:16:15.954 "is_configured": true, 00:16:15.954 "data_offset": 256, 00:16:15.954 "data_size": 7936 00:16:15.954 }, 00:16:15.954 { 00:16:15.954 "name": "BaseBdev2", 00:16:15.954 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:15.954 "is_configured": true, 00:16:15.954 "data_offset": 256, 00:16:15.954 "data_size": 7936 00:16:15.954 } 00:16:15.954 ] 00:16:15.954 }' 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@709 -- # break 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:15.954 "name": "raid_bdev1", 00:16:15.954 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:15.954 "strip_size_kb": 0, 00:16:15.954 "state": "online", 00:16:15.954 "raid_level": "raid1", 00:16:15.954 "superblock": true, 00:16:15.954 "num_base_bdevs": 2, 00:16:15.954 "num_base_bdevs_discovered": 2, 00:16:15.954 "num_base_bdevs_operational": 2, 00:16:15.954 "base_bdevs_list": [ 00:16:15.954 { 00:16:15.954 "name": "spare", 00:16:15.954 "uuid": "284dae40-c263-5a17-a73f-a8e0353939e1", 00:16:15.954 "is_configured": true, 00:16:15.954 "data_offset": 256, 00:16:15.954 "data_size": 7936 00:16:15.954 }, 00:16:15.954 { 00:16:15.954 "name": "BaseBdev2", 00:16:15.954 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:15.954 "is_configured": true, 00:16:15.954 "data_offset": 256, 00:16:15.954 "data_size": 7936 00:16:15.954 } 00:16:15.954 ] 00:16:15.954 }' 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:15.954 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:16.214 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:16.214 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:16.214 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:16.214 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:16.214 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:16.214 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:16.214 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:16.214 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:16.214 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:16.214 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:16.214 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:16.214 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:16.214 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.214 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.215 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:16.215 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.215 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.215 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:16.215 "name": "raid_bdev1", 00:16:16.215 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:16.215 "strip_size_kb": 0, 00:16:16.215 "state": "online", 00:16:16.215 "raid_level": "raid1", 00:16:16.215 "superblock": true, 00:16:16.215 "num_base_bdevs": 2, 00:16:16.215 "num_base_bdevs_discovered": 2, 00:16:16.215 "num_base_bdevs_operational": 2, 00:16:16.215 "base_bdevs_list": [ 00:16:16.215 { 00:16:16.215 "name": "spare", 00:16:16.215 "uuid": "284dae40-c263-5a17-a73f-a8e0353939e1", 00:16:16.215 "is_configured": true, 00:16:16.215 "data_offset": 256, 00:16:16.215 "data_size": 7936 00:16:16.215 }, 00:16:16.215 { 00:16:16.215 "name": "BaseBdev2", 00:16:16.215 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:16.215 "is_configured": true, 00:16:16.215 "data_offset": 256, 00:16:16.215 "data_size": 7936 00:16:16.215 } 00:16:16.215 ] 00:16:16.215 }' 00:16:16.215 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:16.215 19:06:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.785 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.786 [2024-12-05 19:06:34.038394] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:16.786 [2024-12-05 19:06:34.038479] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:16.786 [2024-12-05 19:06:34.038592] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:16.786 [2024-12-05 19:06:34.038708] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:16.786 [2024-12-05 19:06:34.038759] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # jq length 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:16:16.786 /dev/nbd0 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:16.786 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:17.047 1+0 records in 00:16:17.047 1+0 records out 00:16:17.047 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000377505 s, 10.9 MB/s 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:16:17.047 /dev/nbd1 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # local i 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@877 -- # break 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:17.047 1+0 records in 00:16:17.047 1+0 records out 00:16:17.047 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000253145 s, 16.2 MB/s 00:16:17.047 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:17.307 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@890 -- # size=4096 00:16:17.307 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:17.307 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:16:17.307 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@893 -- # return 0 00:16:17.307 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:17.307 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:17.308 19:06:34 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:17.568 [2024-12-05 19:06:35.069160] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:17.568 [2024-12-05 19:06:35.069234] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:17.568 [2024-12-05 19:06:35.069255] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:16:17.568 [2024-12-05 19:06:35.069268] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:17.568 [2024-12-05 19:06:35.071162] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:17.568 [2024-12-05 19:06:35.071202] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:17.568 [2024-12-05 19:06:35.071259] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:17.568 [2024-12-05 19:06:35.071305] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:17.568 [2024-12-05 19:06:35.071430] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:17.568 spare 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:17.568 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:17.828 [2024-12-05 19:06:35.171324] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:16:17.828 [2024-12-05 19:06:35.171349] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:17.828 [2024-12-05 19:06:35.171432] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb1b0 00:16:17.828 [2024-12-05 19:06:35.171528] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:16:17.828 [2024-12-05 19:06:35.171538] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:16:17.828 [2024-12-05 19:06:35.171640] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:17.828 "name": "raid_bdev1", 00:16:17.828 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:17.828 "strip_size_kb": 0, 00:16:17.828 "state": "online", 00:16:17.828 "raid_level": "raid1", 00:16:17.828 "superblock": true, 00:16:17.828 "num_base_bdevs": 2, 00:16:17.828 "num_base_bdevs_discovered": 2, 00:16:17.828 "num_base_bdevs_operational": 2, 00:16:17.828 "base_bdevs_list": [ 00:16:17.828 { 00:16:17.828 "name": "spare", 00:16:17.828 "uuid": "284dae40-c263-5a17-a73f-a8e0353939e1", 00:16:17.828 "is_configured": true, 00:16:17.828 "data_offset": 256, 00:16:17.828 "data_size": 7936 00:16:17.828 }, 00:16:17.828 { 00:16:17.828 "name": "BaseBdev2", 00:16:17.828 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:17.828 "is_configured": true, 00:16:17.828 "data_offset": 256, 00:16:17.828 "data_size": 7936 00:16:17.828 } 00:16:17.828 ] 00:16:17.828 }' 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:17.828 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:18.398 "name": "raid_bdev1", 00:16:18.398 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:18.398 "strip_size_kb": 0, 00:16:18.398 "state": "online", 00:16:18.398 "raid_level": "raid1", 00:16:18.398 "superblock": true, 00:16:18.398 "num_base_bdevs": 2, 00:16:18.398 "num_base_bdevs_discovered": 2, 00:16:18.398 "num_base_bdevs_operational": 2, 00:16:18.398 "base_bdevs_list": [ 00:16:18.398 { 00:16:18.398 "name": "spare", 00:16:18.398 "uuid": "284dae40-c263-5a17-a73f-a8e0353939e1", 00:16:18.398 "is_configured": true, 00:16:18.398 "data_offset": 256, 00:16:18.398 "data_size": 7936 00:16:18.398 }, 00:16:18.398 { 00:16:18.398 "name": "BaseBdev2", 00:16:18.398 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:18.398 "is_configured": true, 00:16:18.398 "data_offset": 256, 00:16:18.398 "data_size": 7936 00:16:18.398 } 00:16:18.398 ] 00:16:18.398 }' 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.398 [2024-12-05 19:06:35.827956] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:18.398 "name": "raid_bdev1", 00:16:18.398 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:18.398 "strip_size_kb": 0, 00:16:18.398 "state": "online", 00:16:18.398 "raid_level": "raid1", 00:16:18.398 "superblock": true, 00:16:18.398 "num_base_bdevs": 2, 00:16:18.398 "num_base_bdevs_discovered": 1, 00:16:18.398 "num_base_bdevs_operational": 1, 00:16:18.398 "base_bdevs_list": [ 00:16:18.398 { 00:16:18.398 "name": null, 00:16:18.398 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:18.398 "is_configured": false, 00:16:18.398 "data_offset": 0, 00:16:18.398 "data_size": 7936 00:16:18.398 }, 00:16:18.398 { 00:16:18.398 "name": "BaseBdev2", 00:16:18.398 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:18.398 "is_configured": true, 00:16:18.398 "data_offset": 256, 00:16:18.398 "data_size": 7936 00:16:18.398 } 00:16:18.398 ] 00:16:18.398 }' 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:18.398 19:06:35 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.969 19:06:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:18.969 19:06:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:18.969 19:06:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:18.969 [2024-12-05 19:06:36.303184] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:18.969 [2024-12-05 19:06:36.303413] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:18.969 [2024-12-05 19:06:36.303483] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:18.969 [2024-12-05 19:06:36.303549] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:18.969 [2024-12-05 19:06:36.305957] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb280 00:16:18.969 [2024-12-05 19:06:36.307825] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:18.969 19:06:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:18.969 19:06:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:19.909 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:19.909 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:19.909 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:19.909 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:19.909 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:19.909 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.909 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:19.909 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:19.909 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:19.909 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:19.909 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:19.909 "name": "raid_bdev1", 00:16:19.909 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:19.909 "strip_size_kb": 0, 00:16:19.909 "state": "online", 00:16:19.909 "raid_level": "raid1", 00:16:19.909 "superblock": true, 00:16:19.909 "num_base_bdevs": 2, 00:16:19.910 "num_base_bdevs_discovered": 2, 00:16:19.910 "num_base_bdevs_operational": 2, 00:16:19.910 "process": { 00:16:19.910 "type": "rebuild", 00:16:19.910 "target": "spare", 00:16:19.910 "progress": { 00:16:19.910 "blocks": 2560, 00:16:19.910 "percent": 32 00:16:19.910 } 00:16:19.910 }, 00:16:19.910 "base_bdevs_list": [ 00:16:19.910 { 00:16:19.910 "name": "spare", 00:16:19.910 "uuid": "284dae40-c263-5a17-a73f-a8e0353939e1", 00:16:19.910 "is_configured": true, 00:16:19.910 "data_offset": 256, 00:16:19.910 "data_size": 7936 00:16:19.910 }, 00:16:19.910 { 00:16:19.910 "name": "BaseBdev2", 00:16:19.910 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:19.910 "is_configured": true, 00:16:19.910 "data_offset": 256, 00:16:19.910 "data_size": 7936 00:16:19.910 } 00:16:19.910 ] 00:16:19.910 }' 00:16:19.910 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:19.910 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:19.910 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:19.910 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:19.910 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:19.910 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:19.910 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:20.170 [2024-12-05 19:06:37.466691] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:20.170 [2024-12-05 19:06:37.512655] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:20.170 [2024-12-05 19:06:37.512716] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:20.170 [2024-12-05 19:06:37.512749] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:20.170 [2024-12-05 19:06:37.512757] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.170 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:20.170 "name": "raid_bdev1", 00:16:20.170 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:20.170 "strip_size_kb": 0, 00:16:20.170 "state": "online", 00:16:20.170 "raid_level": "raid1", 00:16:20.170 "superblock": true, 00:16:20.170 "num_base_bdevs": 2, 00:16:20.170 "num_base_bdevs_discovered": 1, 00:16:20.170 "num_base_bdevs_operational": 1, 00:16:20.170 "base_bdevs_list": [ 00:16:20.170 { 00:16:20.170 "name": null, 00:16:20.170 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:20.170 "is_configured": false, 00:16:20.170 "data_offset": 0, 00:16:20.170 "data_size": 7936 00:16:20.170 }, 00:16:20.170 { 00:16:20.170 "name": "BaseBdev2", 00:16:20.170 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:20.170 "is_configured": true, 00:16:20.170 "data_offset": 256, 00:16:20.171 "data_size": 7936 00:16:20.171 } 00:16:20.171 ] 00:16:20.171 }' 00:16:20.171 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:20.171 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:20.431 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:20.431 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:20.431 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:20.692 [2024-12-05 19:06:37.991067] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:20.692 [2024-12-05 19:06:37.991209] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:20.692 [2024-12-05 19:06:37.991255] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:16:20.692 [2024-12-05 19:06:37.991286] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:20.692 [2024-12-05 19:06:37.991522] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:20.692 [2024-12-05 19:06:37.991571] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:20.692 [2024-12-05 19:06:37.991649] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:20.692 [2024-12-05 19:06:37.991701] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:20.692 [2024-12-05 19:06:37.991743] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:20.692 [2024-12-05 19:06:37.991810] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:20.692 [2024-12-05 19:06:37.994128] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb350 00:16:20.692 [2024-12-05 19:06:37.996022] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:20.692 spare 00:16:20.692 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:20.692 19:06:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:21.633 "name": "raid_bdev1", 00:16:21.633 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:21.633 "strip_size_kb": 0, 00:16:21.633 "state": "online", 00:16:21.633 "raid_level": "raid1", 00:16:21.633 "superblock": true, 00:16:21.633 "num_base_bdevs": 2, 00:16:21.633 "num_base_bdevs_discovered": 2, 00:16:21.633 "num_base_bdevs_operational": 2, 00:16:21.633 "process": { 00:16:21.633 "type": "rebuild", 00:16:21.633 "target": "spare", 00:16:21.633 "progress": { 00:16:21.633 "blocks": 2560, 00:16:21.633 "percent": 32 00:16:21.633 } 00:16:21.633 }, 00:16:21.633 "base_bdevs_list": [ 00:16:21.633 { 00:16:21.633 "name": "spare", 00:16:21.633 "uuid": "284dae40-c263-5a17-a73f-a8e0353939e1", 00:16:21.633 "is_configured": true, 00:16:21.633 "data_offset": 256, 00:16:21.633 "data_size": 7936 00:16:21.633 }, 00:16:21.633 { 00:16:21.633 "name": "BaseBdev2", 00:16:21.633 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:21.633 "is_configured": true, 00:16:21.633 "data_offset": 256, 00:16:21.633 "data_size": 7936 00:16:21.633 } 00:16:21.633 ] 00:16:21.633 }' 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:21.633 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:21.633 [2024-12-05 19:06:39.158830] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:21.892 [2024-12-05 19:06:39.200099] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:21.892 [2024-12-05 19:06:39.200225] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:21.892 [2024-12-05 19:06:39.200261] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:21.892 [2024-12-05 19:06:39.200284] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:21.893 "name": "raid_bdev1", 00:16:21.893 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:21.893 "strip_size_kb": 0, 00:16:21.893 "state": "online", 00:16:21.893 "raid_level": "raid1", 00:16:21.893 "superblock": true, 00:16:21.893 "num_base_bdevs": 2, 00:16:21.893 "num_base_bdevs_discovered": 1, 00:16:21.893 "num_base_bdevs_operational": 1, 00:16:21.893 "base_bdevs_list": [ 00:16:21.893 { 00:16:21.893 "name": null, 00:16:21.893 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:21.893 "is_configured": false, 00:16:21.893 "data_offset": 0, 00:16:21.893 "data_size": 7936 00:16:21.893 }, 00:16:21.893 { 00:16:21.893 "name": "BaseBdev2", 00:16:21.893 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:21.893 "is_configured": true, 00:16:21.893 "data_offset": 256, 00:16:21.893 "data_size": 7936 00:16:21.893 } 00:16:21.893 ] 00:16:21.893 }' 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:21.893 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:22.153 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:22.154 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:22.154 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:22.154 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:22.154 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:22.154 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:22.154 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:22.154 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:22.154 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:22.154 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:22.154 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:22.154 "name": "raid_bdev1", 00:16:22.154 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:22.154 "strip_size_kb": 0, 00:16:22.154 "state": "online", 00:16:22.154 "raid_level": "raid1", 00:16:22.154 "superblock": true, 00:16:22.154 "num_base_bdevs": 2, 00:16:22.154 "num_base_bdevs_discovered": 1, 00:16:22.154 "num_base_bdevs_operational": 1, 00:16:22.154 "base_bdevs_list": [ 00:16:22.154 { 00:16:22.154 "name": null, 00:16:22.154 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:22.154 "is_configured": false, 00:16:22.154 "data_offset": 0, 00:16:22.154 "data_size": 7936 00:16:22.154 }, 00:16:22.154 { 00:16:22.154 "name": "BaseBdev2", 00:16:22.154 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:22.154 "is_configured": true, 00:16:22.154 "data_offset": 256, 00:16:22.154 "data_size": 7936 00:16:22.154 } 00:16:22.154 ] 00:16:22.154 }' 00:16:22.154 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:22.414 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:22.414 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:22.414 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:22.414 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:16:22.414 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:22.414 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:22.414 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:22.414 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:22.414 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:22.414 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:22.414 [2024-12-05 19:06:39.758442] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:22.414 [2024-12-05 19:06:39.758497] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:22.414 [2024-12-05 19:06:39.758519] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:16:22.414 [2024-12-05 19:06:39.758530] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:22.414 [2024-12-05 19:06:39.758722] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:22.414 [2024-12-05 19:06:39.758740] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:22.414 [2024-12-05 19:06:39.758785] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:16:22.414 [2024-12-05 19:06:39.758801] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:22.414 [2024-12-05 19:06:39.758814] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:22.414 [2024-12-05 19:06:39.758832] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:16:22.415 BaseBdev1 00:16:22.415 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:22.415 19:06:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@775 -- # sleep 1 00:16:23.354 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:23.354 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:23.354 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:23.354 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:23.355 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:23.355 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:23.355 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:23.355 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:23.355 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:23.355 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:23.355 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:23.355 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:23.355 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.355 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:23.355 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.355 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:23.355 "name": "raid_bdev1", 00:16:23.355 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:23.355 "strip_size_kb": 0, 00:16:23.355 "state": "online", 00:16:23.355 "raid_level": "raid1", 00:16:23.355 "superblock": true, 00:16:23.355 "num_base_bdevs": 2, 00:16:23.355 "num_base_bdevs_discovered": 1, 00:16:23.355 "num_base_bdevs_operational": 1, 00:16:23.355 "base_bdevs_list": [ 00:16:23.355 { 00:16:23.355 "name": null, 00:16:23.355 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:23.355 "is_configured": false, 00:16:23.355 "data_offset": 0, 00:16:23.355 "data_size": 7936 00:16:23.355 }, 00:16:23.355 { 00:16:23.355 "name": "BaseBdev2", 00:16:23.355 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:23.355 "is_configured": true, 00:16:23.355 "data_offset": 256, 00:16:23.355 "data_size": 7936 00:16:23.355 } 00:16:23.355 ] 00:16:23.355 }' 00:16:23.355 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:23.355 19:06:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:23.924 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:23.924 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:23.924 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:23.924 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:23.924 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:23.924 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:23.924 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:23.924 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.924 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:23.924 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:23.924 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:23.924 "name": "raid_bdev1", 00:16:23.924 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:23.924 "strip_size_kb": 0, 00:16:23.924 "state": "online", 00:16:23.924 "raid_level": "raid1", 00:16:23.924 "superblock": true, 00:16:23.924 "num_base_bdevs": 2, 00:16:23.924 "num_base_bdevs_discovered": 1, 00:16:23.924 "num_base_bdevs_operational": 1, 00:16:23.924 "base_bdevs_list": [ 00:16:23.924 { 00:16:23.924 "name": null, 00:16:23.924 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:23.924 "is_configured": false, 00:16:23.924 "data_offset": 0, 00:16:23.925 "data_size": 7936 00:16:23.925 }, 00:16:23.925 { 00:16:23.925 "name": "BaseBdev2", 00:16:23.925 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:23.925 "is_configured": true, 00:16:23.925 "data_offset": 256, 00:16:23.925 "data_size": 7936 00:16:23.925 } 00:16:23.925 ] 00:16:23.925 }' 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@652 -- # local es=0 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:23.925 [2024-12-05 19:06:41.355861] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:23.925 [2024-12-05 19:06:41.356035] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:23.925 [2024-12-05 19:06:41.356049] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:23.925 request: 00:16:23.925 { 00:16:23.925 "base_bdev": "BaseBdev1", 00:16:23.925 "raid_bdev": "raid_bdev1", 00:16:23.925 "method": "bdev_raid_add_base_bdev", 00:16:23.925 "req_id": 1 00:16:23.925 } 00:16:23.925 Got JSON-RPC error response 00:16:23.925 response: 00:16:23.925 { 00:16:23.925 "code": -22, 00:16:23.925 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:23.925 } 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@655 -- # es=1 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:16:23.925 19:06:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:24.866 "name": "raid_bdev1", 00:16:24.866 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:24.866 "strip_size_kb": 0, 00:16:24.866 "state": "online", 00:16:24.866 "raid_level": "raid1", 00:16:24.866 "superblock": true, 00:16:24.866 "num_base_bdevs": 2, 00:16:24.866 "num_base_bdevs_discovered": 1, 00:16:24.866 "num_base_bdevs_operational": 1, 00:16:24.866 "base_bdevs_list": [ 00:16:24.866 { 00:16:24.866 "name": null, 00:16:24.866 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:24.866 "is_configured": false, 00:16:24.866 "data_offset": 0, 00:16:24.866 "data_size": 7936 00:16:24.866 }, 00:16:24.866 { 00:16:24.866 "name": "BaseBdev2", 00:16:24.866 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:24.866 "is_configured": true, 00:16:24.866 "data_offset": 256, 00:16:24.866 "data_size": 7936 00:16:24.866 } 00:16:24.866 ] 00:16:24.866 }' 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:24.866 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:25.437 "name": "raid_bdev1", 00:16:25.437 "uuid": "75341601-6134-463a-a131-78e8b633cdcb", 00:16:25.437 "strip_size_kb": 0, 00:16:25.437 "state": "online", 00:16:25.437 "raid_level": "raid1", 00:16:25.437 "superblock": true, 00:16:25.437 "num_base_bdevs": 2, 00:16:25.437 "num_base_bdevs_discovered": 1, 00:16:25.437 "num_base_bdevs_operational": 1, 00:16:25.437 "base_bdevs_list": [ 00:16:25.437 { 00:16:25.437 "name": null, 00:16:25.437 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:25.437 "is_configured": false, 00:16:25.437 "data_offset": 0, 00:16:25.437 "data_size": 7936 00:16:25.437 }, 00:16:25.437 { 00:16:25.437 "name": "BaseBdev2", 00:16:25.437 "uuid": "edeb5e3e-4a53-54a4-90cf-adfe7ee44a68", 00:16:25.437 "is_configured": true, 00:16:25.437 "data_offset": 256, 00:16:25.437 "data_size": 7936 00:16:25.437 } 00:16:25.437 ] 00:16:25.437 }' 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@784 -- # killprocess 95067 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@954 -- # '[' -z 95067 ']' 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@958 -- # kill -0 95067 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@959 -- # uname 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:25.437 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 95067 00:16:25.697 killing process with pid 95067 00:16:25.697 Received shutdown signal, test time was about 60.000000 seconds 00:16:25.697 00:16:25.697 Latency(us) 00:16:25.697 [2024-12-05T19:06:43.256Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:25.697 [2024-12-05T19:06:43.257Z] =================================================================================================================== 00:16:25.698 [2024-12-05T19:06:43.257Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:25.698 19:06:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:25.698 19:06:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:25.698 19:06:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@972 -- # echo 'killing process with pid 95067' 00:16:25.698 19:06:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@973 -- # kill 95067 00:16:25.698 [2024-12-05 19:06:43.001839] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:25.698 [2024-12-05 19:06:43.001944] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:25.698 [2024-12-05 19:06:43.001993] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:25.698 [2024-12-05 19:06:43.002001] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:16:25.698 19:06:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@978 -- # wait 95067 00:16:25.698 [2024-12-05 19:06:43.035329] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:25.698 ************************************ 00:16:25.698 END TEST raid_rebuild_test_sb_md_separate 00:16:25.698 ************************************ 00:16:25.698 19:06:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@786 -- # return 0 00:16:25.698 00:16:25.698 real 0m18.372s 00:16:25.698 user 0m24.509s 00:16:25.698 sys 0m2.644s 00:16:25.698 19:06:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:25.698 19:06:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:25.958 19:06:43 bdev_raid -- bdev/bdev_raid.sh@1010 -- # base_malloc_params='-m 32 -i' 00:16:25.958 19:06:43 bdev_raid -- bdev/bdev_raid.sh@1011 -- # run_test raid_state_function_test_sb_md_interleaved raid_state_function_test raid1 2 true 00:16:25.958 19:06:43 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:16:25.958 19:06:43 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:25.958 19:06:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:25.958 ************************************ 00:16:25.958 START TEST raid_state_function_test_sb_md_interleaved 00:16:25.958 ************************************ 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1129 -- # raid_state_function_test raid1 2 true 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # local strip_size 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:16:25.958 Process raid pid: 95634 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@229 -- # raid_pid=95634 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 95634' 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@231 -- # waitforlisten 95634 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 95634 ']' 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:25.958 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:25.958 19:06:43 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:25.958 [2024-12-05 19:06:43.416181] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:16:25.958 [2024-12-05 19:06:43.416365] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:26.219 [2024-12-05 19:06:43.570749] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:26.219 [2024-12-05 19:06:43.595695] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:26.219 [2024-12-05 19:06:43.638937] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:26.219 [2024-12-05 19:06:43.639051] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:26.790 [2024-12-05 19:06:44.265880] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:26.790 [2024-12-05 19:06:44.265950] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:26.790 [2024-12-05 19:06:44.265976] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:26.790 [2024-12-05 19:06:44.265985] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:26.790 "name": "Existed_Raid", 00:16:26.790 "uuid": "7dea841e-0c5f-40cf-8b37-14b815be8d7b", 00:16:26.790 "strip_size_kb": 0, 00:16:26.790 "state": "configuring", 00:16:26.790 "raid_level": "raid1", 00:16:26.790 "superblock": true, 00:16:26.790 "num_base_bdevs": 2, 00:16:26.790 "num_base_bdevs_discovered": 0, 00:16:26.790 "num_base_bdevs_operational": 2, 00:16:26.790 "base_bdevs_list": [ 00:16:26.790 { 00:16:26.790 "name": "BaseBdev1", 00:16:26.790 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:26.790 "is_configured": false, 00:16:26.790 "data_offset": 0, 00:16:26.790 "data_size": 0 00:16:26.790 }, 00:16:26.790 { 00:16:26.790 "name": "BaseBdev2", 00:16:26.790 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:26.790 "is_configured": false, 00:16:26.790 "data_offset": 0, 00:16:26.790 "data_size": 0 00:16:26.790 } 00:16:26.790 ] 00:16:26.790 }' 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:26.790 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:27.361 [2024-12-05 19:06:44.717141] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:27.361 [2024-12-05 19:06:44.717251] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:27.361 [2024-12-05 19:06:44.729010] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:27.361 [2024-12-05 19:06:44.729112] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:27.361 [2024-12-05 19:06:44.729139] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:27.361 [2024-12-05 19:06:44.729173] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:27.361 [2024-12-05 19:06:44.749904] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:27.361 BaseBdev1 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev1 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # local i 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:27.361 [ 00:16:27.361 { 00:16:27.361 "name": "BaseBdev1", 00:16:27.361 "aliases": [ 00:16:27.361 "49f892dc-1bee-41fe-b29b-7499a0793140" 00:16:27.361 ], 00:16:27.361 "product_name": "Malloc disk", 00:16:27.361 "block_size": 4128, 00:16:27.361 "num_blocks": 8192, 00:16:27.361 "uuid": "49f892dc-1bee-41fe-b29b-7499a0793140", 00:16:27.361 "md_size": 32, 00:16:27.361 "md_interleave": true, 00:16:27.361 "dif_type": 0, 00:16:27.361 "assigned_rate_limits": { 00:16:27.361 "rw_ios_per_sec": 0, 00:16:27.361 "rw_mbytes_per_sec": 0, 00:16:27.361 "r_mbytes_per_sec": 0, 00:16:27.361 "w_mbytes_per_sec": 0 00:16:27.361 }, 00:16:27.361 "claimed": true, 00:16:27.361 "claim_type": "exclusive_write", 00:16:27.361 "zoned": false, 00:16:27.361 "supported_io_types": { 00:16:27.361 "read": true, 00:16:27.361 "write": true, 00:16:27.361 "unmap": true, 00:16:27.361 "flush": true, 00:16:27.361 "reset": true, 00:16:27.361 "nvme_admin": false, 00:16:27.361 "nvme_io": false, 00:16:27.361 "nvme_io_md": false, 00:16:27.361 "write_zeroes": true, 00:16:27.361 "zcopy": true, 00:16:27.361 "get_zone_info": false, 00:16:27.361 "zone_management": false, 00:16:27.361 "zone_append": false, 00:16:27.361 "compare": false, 00:16:27.361 "compare_and_write": false, 00:16:27.361 "abort": true, 00:16:27.361 "seek_hole": false, 00:16:27.361 "seek_data": false, 00:16:27.361 "copy": true, 00:16:27.361 "nvme_iov_md": false 00:16:27.361 }, 00:16:27.361 "memory_domains": [ 00:16:27.361 { 00:16:27.361 "dma_device_id": "system", 00:16:27.361 "dma_device_type": 1 00:16:27.361 }, 00:16:27.361 { 00:16:27.361 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:27.361 "dma_device_type": 2 00:16:27.361 } 00:16:27.361 ], 00:16:27.361 "driver_specific": {} 00:16:27.361 } 00:16:27.361 ] 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@911 -- # return 0 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:27.361 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:27.361 "name": "Existed_Raid", 00:16:27.361 "uuid": "4fe68ecb-f784-44d5-9af9-c331ac91b7b2", 00:16:27.361 "strip_size_kb": 0, 00:16:27.362 "state": "configuring", 00:16:27.362 "raid_level": "raid1", 00:16:27.362 "superblock": true, 00:16:27.362 "num_base_bdevs": 2, 00:16:27.362 "num_base_bdevs_discovered": 1, 00:16:27.362 "num_base_bdevs_operational": 2, 00:16:27.362 "base_bdevs_list": [ 00:16:27.362 { 00:16:27.362 "name": "BaseBdev1", 00:16:27.362 "uuid": "49f892dc-1bee-41fe-b29b-7499a0793140", 00:16:27.362 "is_configured": true, 00:16:27.362 "data_offset": 256, 00:16:27.362 "data_size": 7936 00:16:27.362 }, 00:16:27.362 { 00:16:27.362 "name": "BaseBdev2", 00:16:27.362 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:27.362 "is_configured": false, 00:16:27.362 "data_offset": 0, 00:16:27.362 "data_size": 0 00:16:27.362 } 00:16:27.362 ] 00:16:27.362 }' 00:16:27.362 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:27.362 19:06:44 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:27.931 [2024-12-05 19:06:45.245113] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:27.931 [2024-12-05 19:06:45.245204] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:27.931 [2024-12-05 19:06:45.257148] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:27.931 [2024-12-05 19:06:45.259015] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:27.931 [2024-12-05 19:06:45.259092] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:27.931 "name": "Existed_Raid", 00:16:27.931 "uuid": "8f2fa162-fe5f-4106-ad3d-40f4d37b7d95", 00:16:27.931 "strip_size_kb": 0, 00:16:27.931 "state": "configuring", 00:16:27.931 "raid_level": "raid1", 00:16:27.931 "superblock": true, 00:16:27.931 "num_base_bdevs": 2, 00:16:27.931 "num_base_bdevs_discovered": 1, 00:16:27.931 "num_base_bdevs_operational": 2, 00:16:27.931 "base_bdevs_list": [ 00:16:27.931 { 00:16:27.931 "name": "BaseBdev1", 00:16:27.931 "uuid": "49f892dc-1bee-41fe-b29b-7499a0793140", 00:16:27.931 "is_configured": true, 00:16:27.931 "data_offset": 256, 00:16:27.931 "data_size": 7936 00:16:27.931 }, 00:16:27.931 { 00:16:27.931 "name": "BaseBdev2", 00:16:27.931 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:27.931 "is_configured": false, 00:16:27.931 "data_offset": 0, 00:16:27.931 "data_size": 0 00:16:27.931 } 00:16:27.931 ] 00:16:27.931 }' 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:27.931 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:28.190 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2 00:16:28.190 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:28.190 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:28.190 [2024-12-05 19:06:45.711515] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:28.190 [2024-12-05 19:06:45.711809] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:16:28.190 [2024-12-05 19:06:45.711827] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:28.190 [2024-12-05 19:06:45.711940] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:28.190 [2024-12-05 19:06:45.712051] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:16:28.190 [2024-12-05 19:06:45.712065] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:16:28.190 [2024-12-05 19:06:45.712123] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:28.190 BaseBdev2 00:16:28.190 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:28.190 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:16:28.191 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@903 -- # local bdev_name=BaseBdev2 00:16:28.191 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:16:28.191 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@905 -- # local i 00:16:28.191 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:16:28.191 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:16:28.191 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@908 -- # rpc_cmd bdev_wait_for_examine 00:16:28.191 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:28.191 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:28.191 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:28.191 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@910 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:28.191 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:28.191 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:28.191 [ 00:16:28.191 { 00:16:28.191 "name": "BaseBdev2", 00:16:28.191 "aliases": [ 00:16:28.191 "a0dd881e-c931-4e85-9060-ca2862234a76" 00:16:28.191 ], 00:16:28.191 "product_name": "Malloc disk", 00:16:28.191 "block_size": 4128, 00:16:28.191 "num_blocks": 8192, 00:16:28.191 "uuid": "a0dd881e-c931-4e85-9060-ca2862234a76", 00:16:28.191 "md_size": 32, 00:16:28.191 "md_interleave": true, 00:16:28.191 "dif_type": 0, 00:16:28.191 "assigned_rate_limits": { 00:16:28.191 "rw_ios_per_sec": 0, 00:16:28.191 "rw_mbytes_per_sec": 0, 00:16:28.191 "r_mbytes_per_sec": 0, 00:16:28.191 "w_mbytes_per_sec": 0 00:16:28.191 }, 00:16:28.191 "claimed": true, 00:16:28.191 "claim_type": "exclusive_write", 00:16:28.191 "zoned": false, 00:16:28.191 "supported_io_types": { 00:16:28.191 "read": true, 00:16:28.191 "write": true, 00:16:28.191 "unmap": true, 00:16:28.191 "flush": true, 00:16:28.191 "reset": true, 00:16:28.191 "nvme_admin": false, 00:16:28.191 "nvme_io": false, 00:16:28.191 "nvme_io_md": false, 00:16:28.191 "write_zeroes": true, 00:16:28.191 "zcopy": true, 00:16:28.191 "get_zone_info": false, 00:16:28.191 "zone_management": false, 00:16:28.191 "zone_append": false, 00:16:28.191 "compare": false, 00:16:28.191 "compare_and_write": false, 00:16:28.191 "abort": true, 00:16:28.191 "seek_hole": false, 00:16:28.191 "seek_data": false, 00:16:28.191 "copy": true, 00:16:28.191 "nvme_iov_md": false 00:16:28.191 }, 00:16:28.191 "memory_domains": [ 00:16:28.191 { 00:16:28.191 "dma_device_id": "system", 00:16:28.191 "dma_device_type": 1 00:16:28.191 }, 00:16:28.191 { 00:16:28.191 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:28.191 "dma_device_type": 2 00:16:28.191 } 00:16:28.450 ], 00:16:28.450 "driver_specific": {} 00:16:28.450 } 00:16:28.450 ] 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@911 -- # return 0 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:28.450 "name": "Existed_Raid", 00:16:28.450 "uuid": "8f2fa162-fe5f-4106-ad3d-40f4d37b7d95", 00:16:28.450 "strip_size_kb": 0, 00:16:28.450 "state": "online", 00:16:28.450 "raid_level": "raid1", 00:16:28.450 "superblock": true, 00:16:28.450 "num_base_bdevs": 2, 00:16:28.450 "num_base_bdevs_discovered": 2, 00:16:28.450 "num_base_bdevs_operational": 2, 00:16:28.450 "base_bdevs_list": [ 00:16:28.450 { 00:16:28.450 "name": "BaseBdev1", 00:16:28.450 "uuid": "49f892dc-1bee-41fe-b29b-7499a0793140", 00:16:28.450 "is_configured": true, 00:16:28.450 "data_offset": 256, 00:16:28.450 "data_size": 7936 00:16:28.450 }, 00:16:28.450 { 00:16:28.450 "name": "BaseBdev2", 00:16:28.450 "uuid": "a0dd881e-c931-4e85-9060-ca2862234a76", 00:16:28.450 "is_configured": true, 00:16:28.450 "data_offset": 256, 00:16:28.450 "data_size": 7936 00:16:28.450 } 00:16:28.450 ] 00:16:28.450 }' 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:28.450 19:06:45 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:28.709 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:16:28.709 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:16:28.710 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:28.710 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:28.710 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:28.710 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:28.710 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:16:28.710 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:28.710 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:28.710 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:28.710 [2024-12-05 19:06:46.230961] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:28.710 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:28.710 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:28.710 "name": "Existed_Raid", 00:16:28.710 "aliases": [ 00:16:28.710 "8f2fa162-fe5f-4106-ad3d-40f4d37b7d95" 00:16:28.710 ], 00:16:28.710 "product_name": "Raid Volume", 00:16:28.710 "block_size": 4128, 00:16:28.710 "num_blocks": 7936, 00:16:28.710 "uuid": "8f2fa162-fe5f-4106-ad3d-40f4d37b7d95", 00:16:28.710 "md_size": 32, 00:16:28.710 "md_interleave": true, 00:16:28.710 "dif_type": 0, 00:16:28.710 "assigned_rate_limits": { 00:16:28.710 "rw_ios_per_sec": 0, 00:16:28.710 "rw_mbytes_per_sec": 0, 00:16:28.710 "r_mbytes_per_sec": 0, 00:16:28.710 "w_mbytes_per_sec": 0 00:16:28.710 }, 00:16:28.710 "claimed": false, 00:16:28.710 "zoned": false, 00:16:28.710 "supported_io_types": { 00:16:28.710 "read": true, 00:16:28.710 "write": true, 00:16:28.710 "unmap": false, 00:16:28.710 "flush": false, 00:16:28.710 "reset": true, 00:16:28.710 "nvme_admin": false, 00:16:28.710 "nvme_io": false, 00:16:28.710 "nvme_io_md": false, 00:16:28.710 "write_zeroes": true, 00:16:28.710 "zcopy": false, 00:16:28.710 "get_zone_info": false, 00:16:28.710 "zone_management": false, 00:16:28.710 "zone_append": false, 00:16:28.710 "compare": false, 00:16:28.710 "compare_and_write": false, 00:16:28.710 "abort": false, 00:16:28.710 "seek_hole": false, 00:16:28.710 "seek_data": false, 00:16:28.710 "copy": false, 00:16:28.710 "nvme_iov_md": false 00:16:28.710 }, 00:16:28.710 "memory_domains": [ 00:16:28.710 { 00:16:28.710 "dma_device_id": "system", 00:16:28.710 "dma_device_type": 1 00:16:28.710 }, 00:16:28.710 { 00:16:28.710 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:28.710 "dma_device_type": 2 00:16:28.710 }, 00:16:28.710 { 00:16:28.710 "dma_device_id": "system", 00:16:28.710 "dma_device_type": 1 00:16:28.710 }, 00:16:28.710 { 00:16:28.710 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:28.710 "dma_device_type": 2 00:16:28.710 } 00:16:28.710 ], 00:16:28.710 "driver_specific": { 00:16:28.710 "raid": { 00:16:28.710 "uuid": "8f2fa162-fe5f-4106-ad3d-40f4d37b7d95", 00:16:28.710 "strip_size_kb": 0, 00:16:28.710 "state": "online", 00:16:28.710 "raid_level": "raid1", 00:16:28.710 "superblock": true, 00:16:28.710 "num_base_bdevs": 2, 00:16:28.710 "num_base_bdevs_discovered": 2, 00:16:28.710 "num_base_bdevs_operational": 2, 00:16:28.710 "base_bdevs_list": [ 00:16:28.710 { 00:16:28.710 "name": "BaseBdev1", 00:16:28.710 "uuid": "49f892dc-1bee-41fe-b29b-7499a0793140", 00:16:28.710 "is_configured": true, 00:16:28.710 "data_offset": 256, 00:16:28.710 "data_size": 7936 00:16:28.710 }, 00:16:28.710 { 00:16:28.710 "name": "BaseBdev2", 00:16:28.710 "uuid": "a0dd881e-c931-4e85-9060-ca2862234a76", 00:16:28.710 "is_configured": true, 00:16:28.710 "data_offset": 256, 00:16:28.710 "data_size": 7936 00:16:28.710 } 00:16:28.710 ] 00:16:28.710 } 00:16:28.710 } 00:16:28.710 }' 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:16:28.970 BaseBdev2' 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:28.970 [2024-12-05 19:06:46.474345] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@260 -- # local expected_state 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:28.970 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:29.230 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:29.230 "name": "Existed_Raid", 00:16:29.230 "uuid": "8f2fa162-fe5f-4106-ad3d-40f4d37b7d95", 00:16:29.230 "strip_size_kb": 0, 00:16:29.230 "state": "online", 00:16:29.230 "raid_level": "raid1", 00:16:29.230 "superblock": true, 00:16:29.230 "num_base_bdevs": 2, 00:16:29.230 "num_base_bdevs_discovered": 1, 00:16:29.230 "num_base_bdevs_operational": 1, 00:16:29.230 "base_bdevs_list": [ 00:16:29.230 { 00:16:29.230 "name": null, 00:16:29.230 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:29.230 "is_configured": false, 00:16:29.230 "data_offset": 0, 00:16:29.230 "data_size": 7936 00:16:29.230 }, 00:16:29.230 { 00:16:29.230 "name": "BaseBdev2", 00:16:29.230 "uuid": "a0dd881e-c931-4e85-9060-ca2862234a76", 00:16:29.230 "is_configured": true, 00:16:29.230 "data_offset": 256, 00:16:29.230 "data_size": 7936 00:16:29.230 } 00:16:29.230 ] 00:16:29.230 }' 00:16:29.230 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:29.230 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:29.490 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:16:29.490 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:29.490 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:29.490 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:16:29.490 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:29.490 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:29.490 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:29.490 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:16:29.490 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:29.490 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:16:29.491 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:29.491 19:06:46 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:29.491 [2024-12-05 19:06:47.005343] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:29.491 [2024-12-05 19:06:47.005509] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:29.491 [2024-12-05 19:06:47.017571] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:29.491 [2024-12-05 19:06:47.017736] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:29.491 [2024-12-05 19:06:47.017794] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:16:29.491 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:29.491 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:16:29.491 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:29.491 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:29.491 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:16:29.491 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:29.491 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:29.491 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:29.751 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:16:29.751 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:16:29.751 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:16:29.751 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@326 -- # killprocess 95634 00:16:29.751 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 95634 ']' 00:16:29.751 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 95634 00:16:29.751 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:16:29.751 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:29.751 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 95634 00:16:29.751 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:29.751 killing process with pid 95634 00:16:29.751 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:29.751 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 95634' 00:16:29.751 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@973 -- # kill 95634 00:16:29.751 [2024-12-05 19:06:47.113969] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:29.751 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@978 -- # wait 95634 00:16:29.751 [2024-12-05 19:06:47.114945] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:30.011 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@328 -- # return 0 00:16:30.011 00:16:30.011 real 0m4.018s 00:16:30.011 user 0m6.331s 00:16:30.011 sys 0m0.904s 00:16:30.011 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:30.011 19:06:47 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.011 ************************************ 00:16:30.011 END TEST raid_state_function_test_sb_md_interleaved 00:16:30.011 ************************************ 00:16:30.011 19:06:47 bdev_raid -- bdev/bdev_raid.sh@1012 -- # run_test raid_superblock_test_md_interleaved raid_superblock_test raid1 2 00:16:30.011 19:06:47 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:16:30.011 19:06:47 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:30.011 19:06:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:30.011 ************************************ 00:16:30.011 START TEST raid_superblock_test_md_interleaved 00:16:30.011 ************************************ 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1129 -- # raid_superblock_test raid1 2 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@399 -- # local strip_size 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@412 -- # raid_pid=95851 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@413 -- # waitforlisten 95851 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 95851 ']' 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:30.011 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:30.011 19:06:47 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.011 [2024-12-05 19:06:47.508063] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:16:30.011 [2024-12-05 19:06:47.508199] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid95851 ] 00:16:30.271 [2024-12-05 19:06:47.664652] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:30.271 [2024-12-05 19:06:47.689790] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:30.271 [2024-12-05 19:06:47.732997] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:30.271 [2024-12-05 19:06:47.733035] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc1 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.841 malloc1 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.841 [2024-12-05 19:06:48.332971] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:30.841 [2024-12-05 19:06:48.333142] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:30.841 [2024-12-05 19:06:48.333188] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:30.841 [2024-12-05 19:06:48.333222] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:30.841 [2024-12-05 19:06:48.335169] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:30.841 [2024-12-05 19:06:48.335245] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:30.841 pt1 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc2 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.841 malloc2 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.841 [2024-12-05 19:06:48.365658] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:30.841 [2024-12-05 19:06:48.365796] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:30.841 [2024-12-05 19:06:48.365830] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:30.841 [2024-12-05 19:06:48.365858] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:30.841 [2024-12-05 19:06:48.367738] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:30.841 [2024-12-05 19:06:48.367809] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:30.841 pt2 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.841 [2024-12-05 19:06:48.377684] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:30.841 [2024-12-05 19:06:48.379559] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:30.841 [2024-12-05 19:06:48.379715] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:30.841 [2024-12-05 19:06:48.379731] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:30.841 [2024-12-05 19:06:48.379834] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:30.841 [2024-12-05 19:06:48.379907] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:30.841 [2024-12-05 19:06:48.379922] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:30.841 [2024-12-05 19:06:48.379989] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:30.841 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.104 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.104 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:31.104 "name": "raid_bdev1", 00:16:31.104 "uuid": "d4059a57-958b-4072-b497-30c60a42e37c", 00:16:31.104 "strip_size_kb": 0, 00:16:31.104 "state": "online", 00:16:31.104 "raid_level": "raid1", 00:16:31.104 "superblock": true, 00:16:31.104 "num_base_bdevs": 2, 00:16:31.104 "num_base_bdevs_discovered": 2, 00:16:31.104 "num_base_bdevs_operational": 2, 00:16:31.104 "base_bdevs_list": [ 00:16:31.104 { 00:16:31.104 "name": "pt1", 00:16:31.104 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:31.104 "is_configured": true, 00:16:31.104 "data_offset": 256, 00:16:31.104 "data_size": 7936 00:16:31.104 }, 00:16:31.104 { 00:16:31.104 "name": "pt2", 00:16:31.104 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:31.104 "is_configured": true, 00:16:31.104 "data_offset": 256, 00:16:31.104 "data_size": 7936 00:16:31.104 } 00:16:31.104 ] 00:16:31.104 }' 00:16:31.104 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:31.104 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.372 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:16:31.372 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:31.372 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:31.372 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:31.372 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:31.372 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:31.372 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:31.372 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:31.372 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.372 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.372 [2024-12-05 19:06:48.841084] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:31.372 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.372 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:31.372 "name": "raid_bdev1", 00:16:31.372 "aliases": [ 00:16:31.372 "d4059a57-958b-4072-b497-30c60a42e37c" 00:16:31.372 ], 00:16:31.372 "product_name": "Raid Volume", 00:16:31.372 "block_size": 4128, 00:16:31.372 "num_blocks": 7936, 00:16:31.372 "uuid": "d4059a57-958b-4072-b497-30c60a42e37c", 00:16:31.372 "md_size": 32, 00:16:31.372 "md_interleave": true, 00:16:31.372 "dif_type": 0, 00:16:31.372 "assigned_rate_limits": { 00:16:31.372 "rw_ios_per_sec": 0, 00:16:31.372 "rw_mbytes_per_sec": 0, 00:16:31.372 "r_mbytes_per_sec": 0, 00:16:31.372 "w_mbytes_per_sec": 0 00:16:31.372 }, 00:16:31.372 "claimed": false, 00:16:31.372 "zoned": false, 00:16:31.372 "supported_io_types": { 00:16:31.372 "read": true, 00:16:31.372 "write": true, 00:16:31.372 "unmap": false, 00:16:31.372 "flush": false, 00:16:31.372 "reset": true, 00:16:31.372 "nvme_admin": false, 00:16:31.372 "nvme_io": false, 00:16:31.372 "nvme_io_md": false, 00:16:31.372 "write_zeroes": true, 00:16:31.372 "zcopy": false, 00:16:31.372 "get_zone_info": false, 00:16:31.372 "zone_management": false, 00:16:31.372 "zone_append": false, 00:16:31.372 "compare": false, 00:16:31.372 "compare_and_write": false, 00:16:31.372 "abort": false, 00:16:31.372 "seek_hole": false, 00:16:31.372 "seek_data": false, 00:16:31.372 "copy": false, 00:16:31.372 "nvme_iov_md": false 00:16:31.372 }, 00:16:31.373 "memory_domains": [ 00:16:31.373 { 00:16:31.373 "dma_device_id": "system", 00:16:31.373 "dma_device_type": 1 00:16:31.373 }, 00:16:31.373 { 00:16:31.373 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:31.373 "dma_device_type": 2 00:16:31.373 }, 00:16:31.373 { 00:16:31.373 "dma_device_id": "system", 00:16:31.373 "dma_device_type": 1 00:16:31.373 }, 00:16:31.373 { 00:16:31.373 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:31.373 "dma_device_type": 2 00:16:31.373 } 00:16:31.373 ], 00:16:31.373 "driver_specific": { 00:16:31.373 "raid": { 00:16:31.373 "uuid": "d4059a57-958b-4072-b497-30c60a42e37c", 00:16:31.373 "strip_size_kb": 0, 00:16:31.373 "state": "online", 00:16:31.373 "raid_level": "raid1", 00:16:31.373 "superblock": true, 00:16:31.373 "num_base_bdevs": 2, 00:16:31.373 "num_base_bdevs_discovered": 2, 00:16:31.373 "num_base_bdevs_operational": 2, 00:16:31.373 "base_bdevs_list": [ 00:16:31.373 { 00:16:31.373 "name": "pt1", 00:16:31.373 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:31.373 "is_configured": true, 00:16:31.373 "data_offset": 256, 00:16:31.373 "data_size": 7936 00:16:31.373 }, 00:16:31.373 { 00:16:31.373 "name": "pt2", 00:16:31.373 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:31.373 "is_configured": true, 00:16:31.373 "data_offset": 256, 00:16:31.373 "data_size": 7936 00:16:31.373 } 00:16:31.373 ] 00:16:31.373 } 00:16:31.373 } 00:16:31.373 }' 00:16:31.373 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:31.373 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:31.373 pt2' 00:16:31.373 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:31.652 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:31.652 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:31.652 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:31.652 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:31.652 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.652 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.652 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.652 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:31.652 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:31.652 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:31.652 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:31.652 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:31.652 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.652 19:06:48 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.652 [2024-12-05 19:06:49.036803] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=d4059a57-958b-4072-b497-30c60a42e37c 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@436 -- # '[' -z d4059a57-958b-4072-b497-30c60a42e37c ']' 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.652 [2024-12-05 19:06:49.076511] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:31.652 [2024-12-05 19:06:49.076538] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:31.652 [2024-12-05 19:06:49.076603] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:31.652 [2024-12-05 19:06:49.076666] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:31.652 [2024-12-05 19:06:49.076694] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.652 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@652 -- # local es=0 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.929 [2024-12-05 19:06:49.212272] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:16:31.929 [2024-12-05 19:06:49.214102] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:16:31.929 [2024-12-05 19:06:49.214164] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:16:31.929 [2024-12-05 19:06:49.214222] bdev_raid.c:3233:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:16:31.929 [2024-12-05 19:06:49.214240] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:31.929 [2024-12-05 19:06:49.214249] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:16:31.929 request: 00:16:31.929 { 00:16:31.929 "name": "raid_bdev1", 00:16:31.929 "raid_level": "raid1", 00:16:31.929 "base_bdevs": [ 00:16:31.929 "malloc1", 00:16:31.929 "malloc2" 00:16:31.929 ], 00:16:31.929 "superblock": false, 00:16:31.929 "method": "bdev_raid_create", 00:16:31.929 "req_id": 1 00:16:31.929 } 00:16:31.929 Got JSON-RPC error response 00:16:31.929 response: 00:16:31.929 { 00:16:31.929 "code": -17, 00:16:31.929 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:16:31.929 } 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@655 -- # es=1 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.929 [2024-12-05 19:06:49.276130] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:31.929 [2024-12-05 19:06:49.276230] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:31.929 [2024-12-05 19:06:49.276280] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:31.929 [2024-12-05 19:06:49.276322] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:31.929 [2024-12-05 19:06:49.278212] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:31.929 [2024-12-05 19:06:49.278280] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:31.929 [2024-12-05 19:06:49.278375] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:31.929 [2024-12-05 19:06:49.278430] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:31.929 pt1 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:31.929 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:31.930 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:31.930 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:31.930 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:31.930 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:31.930 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:31.930 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:31.930 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:31.930 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:31.930 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:31.930 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:31.930 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.930 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:31.930 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:31.930 "name": "raid_bdev1", 00:16:31.930 "uuid": "d4059a57-958b-4072-b497-30c60a42e37c", 00:16:31.930 "strip_size_kb": 0, 00:16:31.930 "state": "configuring", 00:16:31.930 "raid_level": "raid1", 00:16:31.930 "superblock": true, 00:16:31.930 "num_base_bdevs": 2, 00:16:31.930 "num_base_bdevs_discovered": 1, 00:16:31.930 "num_base_bdevs_operational": 2, 00:16:31.930 "base_bdevs_list": [ 00:16:31.930 { 00:16:31.930 "name": "pt1", 00:16:31.930 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:31.930 "is_configured": true, 00:16:31.930 "data_offset": 256, 00:16:31.930 "data_size": 7936 00:16:31.930 }, 00:16:31.930 { 00:16:31.930 "name": null, 00:16:31.930 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:31.930 "is_configured": false, 00:16:31.930 "data_offset": 256, 00:16:31.930 "data_size": 7936 00:16:31.930 } 00:16:31.930 ] 00:16:31.930 }' 00:16:31.930 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:31.930 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.190 [2024-12-05 19:06:49.711500] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:32.190 [2024-12-05 19:06:49.711553] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:32.190 [2024-12-05 19:06:49.711588] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:16:32.190 [2024-12-05 19:06:49.711596] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:32.190 [2024-12-05 19:06:49.711760] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:32.190 [2024-12-05 19:06:49.711774] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:32.190 [2024-12-05 19:06:49.711815] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:32.190 [2024-12-05 19:06:49.711834] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:32.190 [2024-12-05 19:06:49.711906] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:16:32.190 [2024-12-05 19:06:49.711914] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:32.190 [2024-12-05 19:06:49.711981] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:16:32.190 [2024-12-05 19:06:49.712044] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:16:32.190 [2024-12-05 19:06:49.712056] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:16:32.190 [2024-12-05 19:06:49.712107] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:32.190 pt2 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.190 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:32.450 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:32.450 "name": "raid_bdev1", 00:16:32.450 "uuid": "d4059a57-958b-4072-b497-30c60a42e37c", 00:16:32.450 "strip_size_kb": 0, 00:16:32.450 "state": "online", 00:16:32.450 "raid_level": "raid1", 00:16:32.450 "superblock": true, 00:16:32.450 "num_base_bdevs": 2, 00:16:32.450 "num_base_bdevs_discovered": 2, 00:16:32.450 "num_base_bdevs_operational": 2, 00:16:32.450 "base_bdevs_list": [ 00:16:32.450 { 00:16:32.450 "name": "pt1", 00:16:32.450 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:32.450 "is_configured": true, 00:16:32.450 "data_offset": 256, 00:16:32.450 "data_size": 7936 00:16:32.450 }, 00:16:32.450 { 00:16:32.450 "name": "pt2", 00:16:32.450 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:32.450 "is_configured": true, 00:16:32.450 "data_offset": 256, 00:16:32.450 "data_size": 7936 00:16:32.450 } 00:16:32.450 ] 00:16:32.450 }' 00:16:32.450 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:32.450 19:06:49 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.710 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:16:32.710 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:32.710 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:32.710 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:32.710 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:32.710 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:32.710 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:32.710 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:32.710 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.710 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:32.710 [2024-12-05 19:06:50.162981] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:32.710 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:32.710 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:32.710 "name": "raid_bdev1", 00:16:32.710 "aliases": [ 00:16:32.710 "d4059a57-958b-4072-b497-30c60a42e37c" 00:16:32.710 ], 00:16:32.710 "product_name": "Raid Volume", 00:16:32.710 "block_size": 4128, 00:16:32.710 "num_blocks": 7936, 00:16:32.710 "uuid": "d4059a57-958b-4072-b497-30c60a42e37c", 00:16:32.710 "md_size": 32, 00:16:32.710 "md_interleave": true, 00:16:32.710 "dif_type": 0, 00:16:32.710 "assigned_rate_limits": { 00:16:32.710 "rw_ios_per_sec": 0, 00:16:32.710 "rw_mbytes_per_sec": 0, 00:16:32.710 "r_mbytes_per_sec": 0, 00:16:32.710 "w_mbytes_per_sec": 0 00:16:32.710 }, 00:16:32.710 "claimed": false, 00:16:32.710 "zoned": false, 00:16:32.710 "supported_io_types": { 00:16:32.710 "read": true, 00:16:32.710 "write": true, 00:16:32.710 "unmap": false, 00:16:32.710 "flush": false, 00:16:32.710 "reset": true, 00:16:32.710 "nvme_admin": false, 00:16:32.710 "nvme_io": false, 00:16:32.710 "nvme_io_md": false, 00:16:32.710 "write_zeroes": true, 00:16:32.710 "zcopy": false, 00:16:32.710 "get_zone_info": false, 00:16:32.710 "zone_management": false, 00:16:32.710 "zone_append": false, 00:16:32.710 "compare": false, 00:16:32.710 "compare_and_write": false, 00:16:32.710 "abort": false, 00:16:32.710 "seek_hole": false, 00:16:32.710 "seek_data": false, 00:16:32.710 "copy": false, 00:16:32.710 "nvme_iov_md": false 00:16:32.710 }, 00:16:32.710 "memory_domains": [ 00:16:32.710 { 00:16:32.710 "dma_device_id": "system", 00:16:32.710 "dma_device_type": 1 00:16:32.710 }, 00:16:32.710 { 00:16:32.710 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:32.710 "dma_device_type": 2 00:16:32.710 }, 00:16:32.710 { 00:16:32.710 "dma_device_id": "system", 00:16:32.710 "dma_device_type": 1 00:16:32.710 }, 00:16:32.710 { 00:16:32.710 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:32.710 "dma_device_type": 2 00:16:32.710 } 00:16:32.710 ], 00:16:32.710 "driver_specific": { 00:16:32.710 "raid": { 00:16:32.710 "uuid": "d4059a57-958b-4072-b497-30c60a42e37c", 00:16:32.710 "strip_size_kb": 0, 00:16:32.710 "state": "online", 00:16:32.710 "raid_level": "raid1", 00:16:32.710 "superblock": true, 00:16:32.710 "num_base_bdevs": 2, 00:16:32.710 "num_base_bdevs_discovered": 2, 00:16:32.710 "num_base_bdevs_operational": 2, 00:16:32.710 "base_bdevs_list": [ 00:16:32.710 { 00:16:32.711 "name": "pt1", 00:16:32.711 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:32.711 "is_configured": true, 00:16:32.711 "data_offset": 256, 00:16:32.711 "data_size": 7936 00:16:32.711 }, 00:16:32.711 { 00:16:32.711 "name": "pt2", 00:16:32.711 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:32.711 "is_configured": true, 00:16:32.711 "data_offset": 256, 00:16:32.711 "data_size": 7936 00:16:32.711 } 00:16:32.711 ] 00:16:32.711 } 00:16:32.711 } 00:16:32.711 }' 00:16:32.711 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:32.711 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:32.711 pt2' 00:16:32.711 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.971 [2024-12-05 19:06:50.414512] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # '[' d4059a57-958b-4072-b497-30c60a42e37c '!=' d4059a57-958b-4072-b497-30c60a42e37c ']' 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.971 [2024-12-05 19:06:50.462229] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:32.971 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:32.971 "name": "raid_bdev1", 00:16:32.971 "uuid": "d4059a57-958b-4072-b497-30c60a42e37c", 00:16:32.971 "strip_size_kb": 0, 00:16:32.972 "state": "online", 00:16:32.972 "raid_level": "raid1", 00:16:32.972 "superblock": true, 00:16:32.972 "num_base_bdevs": 2, 00:16:32.972 "num_base_bdevs_discovered": 1, 00:16:32.972 "num_base_bdevs_operational": 1, 00:16:32.972 "base_bdevs_list": [ 00:16:32.972 { 00:16:32.972 "name": null, 00:16:32.972 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:32.972 "is_configured": false, 00:16:32.972 "data_offset": 0, 00:16:32.972 "data_size": 7936 00:16:32.972 }, 00:16:32.972 { 00:16:32.972 "name": "pt2", 00:16:32.972 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:32.972 "is_configured": true, 00:16:32.972 "data_offset": 256, 00:16:32.972 "data_size": 7936 00:16:32.972 } 00:16:32.972 ] 00:16:32.972 }' 00:16:32.972 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:32.972 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:33.543 [2024-12-05 19:06:50.925419] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:33.543 [2024-12-05 19:06:50.925445] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:33.543 [2024-12-05 19:06:50.925500] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:33.543 [2024-12-05 19:06:50.925542] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:33.543 [2024-12-05 19:06:50.925551] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@519 -- # i=1 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:33.543 19:06:50 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:33.543 [2024-12-05 19:06:51.001296] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:33.543 [2024-12-05 19:06:51.001407] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:33.543 [2024-12-05 19:06:51.001428] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:16:33.543 [2024-12-05 19:06:51.001437] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:33.543 [2024-12-05 19:06:51.003340] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:33.543 [2024-12-05 19:06:51.003378] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:33.543 [2024-12-05 19:06:51.003424] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:33.543 [2024-12-05 19:06:51.003458] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:33.543 [2024-12-05 19:06:51.003519] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:16:33.543 [2024-12-05 19:06:51.003527] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:33.543 [2024-12-05 19:06:51.003590] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:16:33.543 [2024-12-05 19:06:51.003643] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:16:33.543 [2024-12-05 19:06:51.003652] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:16:33.543 [2024-12-05 19:06:51.003716] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:33.543 pt2 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:33.543 "name": "raid_bdev1", 00:16:33.543 "uuid": "d4059a57-958b-4072-b497-30c60a42e37c", 00:16:33.543 "strip_size_kb": 0, 00:16:33.543 "state": "online", 00:16:33.543 "raid_level": "raid1", 00:16:33.543 "superblock": true, 00:16:33.543 "num_base_bdevs": 2, 00:16:33.543 "num_base_bdevs_discovered": 1, 00:16:33.543 "num_base_bdevs_operational": 1, 00:16:33.543 "base_bdevs_list": [ 00:16:33.543 { 00:16:33.543 "name": null, 00:16:33.543 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:33.543 "is_configured": false, 00:16:33.543 "data_offset": 256, 00:16:33.543 "data_size": 7936 00:16:33.543 }, 00:16:33.543 { 00:16:33.543 "name": "pt2", 00:16:33.543 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:33.543 "is_configured": true, 00:16:33.543 "data_offset": 256, 00:16:33.543 "data_size": 7936 00:16:33.543 } 00:16:33.543 ] 00:16:33.543 }' 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:33.543 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:34.114 [2024-12-05 19:06:51.420613] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:34.114 [2024-12-05 19:06:51.420639] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:34.114 [2024-12-05 19:06:51.420714] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:34.114 [2024-12-05 19:06:51.420756] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:34.114 [2024-12-05 19:06:51.420769] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:34.114 [2024-12-05 19:06:51.480519] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:34.114 [2024-12-05 19:06:51.480635] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:34.114 [2024-12-05 19:06:51.480690] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:16:34.114 [2024-12-05 19:06:51.480723] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:34.114 [2024-12-05 19:06:51.482591] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:34.114 [2024-12-05 19:06:51.482689] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:34.114 [2024-12-05 19:06:51.482787] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:34.114 [2024-12-05 19:06:51.482845] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:34.114 [2024-12-05 19:06:51.482959] bdev_raid.c:3685:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:34.114 [2024-12-05 19:06:51.483011] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:34.114 [2024-12-05 19:06:51.483061] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:16:34.114 [2024-12-05 19:06:51.483130] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:34.114 [2024-12-05 19:06:51.483212] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:16:34.114 [2024-12-05 19:06:51.483253] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:34.114 [2024-12-05 19:06:51.483361] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:34.114 [2024-12-05 19:06:51.483424] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:16:34.114 [2024-12-05 19:06:51.483431] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:16:34.114 [2024-12-05 19:06:51.483493] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:34.114 pt1 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:34.114 "name": "raid_bdev1", 00:16:34.114 "uuid": "d4059a57-958b-4072-b497-30c60a42e37c", 00:16:34.114 "strip_size_kb": 0, 00:16:34.114 "state": "online", 00:16:34.114 "raid_level": "raid1", 00:16:34.114 "superblock": true, 00:16:34.114 "num_base_bdevs": 2, 00:16:34.114 "num_base_bdevs_discovered": 1, 00:16:34.114 "num_base_bdevs_operational": 1, 00:16:34.114 "base_bdevs_list": [ 00:16:34.114 { 00:16:34.114 "name": null, 00:16:34.114 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:34.114 "is_configured": false, 00:16:34.114 "data_offset": 256, 00:16:34.114 "data_size": 7936 00:16:34.114 }, 00:16:34.114 { 00:16:34.114 "name": "pt2", 00:16:34.114 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:34.114 "is_configured": true, 00:16:34.114 "data_offset": 256, 00:16:34.114 "data_size": 7936 00:16:34.114 } 00:16:34.114 ] 00:16:34.114 }' 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:34.114 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:34.374 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:16:34.374 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:34.374 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:34.374 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:34.634 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:34.634 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:16:34.634 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:34.634 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:16:34.634 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:34.634 19:06:51 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:34.634 [2024-12-05 19:06:51.983905] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:34.634 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:34.634 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # '[' d4059a57-958b-4072-b497-30c60a42e37c '!=' d4059a57-958b-4072-b497-30c60a42e37c ']' 00:16:34.634 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@563 -- # killprocess 95851 00:16:34.634 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 95851 ']' 00:16:34.634 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 95851 00:16:34.634 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:16:34.634 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:34.634 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 95851 00:16:34.634 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:34.634 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:34.634 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 95851' 00:16:34.634 killing process with pid 95851 00:16:34.634 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@973 -- # kill 95851 00:16:34.634 [2024-12-05 19:06:52.055525] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:34.634 [2024-12-05 19:06:52.055648] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:34.634 [2024-12-05 19:06:52.055736] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to fr 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@978 -- # wait 95851 00:16:34.634 ee all in destruct 00:16:34.635 [2024-12-05 19:06:52.055803] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:16:34.635 [2024-12-05 19:06:52.079102] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:34.895 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@565 -- # return 0 00:16:34.895 00:16:34.895 real 0m4.880s 00:16:34.895 user 0m7.976s 00:16:34.895 sys 0m1.115s 00:16:34.895 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:34.895 19:06:52 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:34.895 ************************************ 00:16:34.895 END TEST raid_superblock_test_md_interleaved 00:16:34.895 ************************************ 00:16:34.895 19:06:52 bdev_raid -- bdev/bdev_raid.sh@1013 -- # run_test raid_rebuild_test_sb_md_interleaved raid_rebuild_test raid1 2 true false false 00:16:34.895 19:06:52 bdev_raid -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:16:34.895 19:06:52 bdev_raid -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:34.895 19:06:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:34.895 ************************************ 00:16:34.895 START TEST raid_rebuild_test_sb_md_interleaved 00:16:34.895 ************************************ 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1129 -- # raid_rebuild_test raid1 2 true false false 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # local verify=false 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@597 -- # raid_pid=96137 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@598 -- # waitforlisten 96137 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # '[' -z 96137 ']' 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:34.895 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:34.895 19:06:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:35.155 [2024-12-05 19:06:52.470493] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:16:35.155 I/O size of 3145728 is greater than zero copy threshold (65536). 00:16:35.155 Zero copy mechanism will not be used. 00:16:35.155 [2024-12-05 19:06:52.470719] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96137 ] 00:16:35.155 [2024-12-05 19:06:52.622256] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:35.155 [2024-12-05 19:06:52.647530] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:35.155 [2024-12-05 19:06:52.690937] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:35.155 [2024-12-05 19:06:52.691065] bdev_raid.c:1456:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@868 -- # return 0 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1_malloc 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.098 BaseBdev1_malloc 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.098 [2024-12-05 19:06:53.307033] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:36.098 [2024-12-05 19:06:53.307122] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:36.098 [2024-12-05 19:06:53.307155] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:36.098 [2024-12-05 19:06:53.307166] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:36.098 [2024-12-05 19:06:53.309065] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:36.098 [2024-12-05 19:06:53.309102] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:36.098 BaseBdev1 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2_malloc 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.098 BaseBdev2_malloc 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.098 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.098 [2024-12-05 19:06:53.335711] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:36.098 [2024-12-05 19:06:53.335761] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:36.098 [2024-12-05 19:06:53.335796] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:36.098 [2024-12-05 19:06:53.335805] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:36.098 [2024-12-05 19:06:53.337614] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:36.099 [2024-12-05 19:06:53.337655] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:36.099 BaseBdev2 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b spare_malloc 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.099 spare_malloc 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.099 spare_delay 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.099 [2024-12-05 19:06:53.397015] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:36.099 [2024-12-05 19:06:53.397091] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:36.099 [2024-12-05 19:06:53.397124] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:16:36.099 [2024-12-05 19:06:53.397138] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:36.099 [2024-12-05 19:06:53.400119] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:36.099 [2024-12-05 19:06:53.400171] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:36.099 spare 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.099 [2024-12-05 19:06:53.409029] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:36.099 [2024-12-05 19:06:53.411136] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:36.099 [2024-12-05 19:06:53.411404] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:36.099 [2024-12-05 19:06:53.411425] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:36.099 [2024-12-05 19:06:53.411524] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:16:36.099 [2024-12-05 19:06:53.411607] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:36.099 [2024-12-05 19:06:53.411625] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:36.099 [2024-12-05 19:06:53.411723] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:36.099 "name": "raid_bdev1", 00:16:36.099 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:36.099 "strip_size_kb": 0, 00:16:36.099 "state": "online", 00:16:36.099 "raid_level": "raid1", 00:16:36.099 "superblock": true, 00:16:36.099 "num_base_bdevs": 2, 00:16:36.099 "num_base_bdevs_discovered": 2, 00:16:36.099 "num_base_bdevs_operational": 2, 00:16:36.099 "base_bdevs_list": [ 00:16:36.099 { 00:16:36.099 "name": "BaseBdev1", 00:16:36.099 "uuid": "b6c50f4a-2966-5810-a0f3-6c3faba17b5b", 00:16:36.099 "is_configured": true, 00:16:36.099 "data_offset": 256, 00:16:36.099 "data_size": 7936 00:16:36.099 }, 00:16:36.099 { 00:16:36.099 "name": "BaseBdev2", 00:16:36.099 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:36.099 "is_configured": true, 00:16:36.099 "data_offset": 256, 00:16:36.099 "data_size": 7936 00:16:36.099 } 00:16:36.099 ] 00:16:36.099 }' 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:36.099 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.359 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:36.359 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:36.359 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.359 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.359 [2024-12-05 19:06:53.844486] bdev_raid.c:1133:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:36.359 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.359 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:16:36.359 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:36.359 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.359 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:36.359 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.359 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.618 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:16:36.618 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:36.618 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@624 -- # '[' false = true ']' 00:16:36.618 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:36.618 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.618 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.618 [2024-12-05 19:06:53.928116] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:36.619 "name": "raid_bdev1", 00:16:36.619 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:36.619 "strip_size_kb": 0, 00:16:36.619 "state": "online", 00:16:36.619 "raid_level": "raid1", 00:16:36.619 "superblock": true, 00:16:36.619 "num_base_bdevs": 2, 00:16:36.619 "num_base_bdevs_discovered": 1, 00:16:36.619 "num_base_bdevs_operational": 1, 00:16:36.619 "base_bdevs_list": [ 00:16:36.619 { 00:16:36.619 "name": null, 00:16:36.619 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:36.619 "is_configured": false, 00:16:36.619 "data_offset": 0, 00:16:36.619 "data_size": 7936 00:16:36.619 }, 00:16:36.619 { 00:16:36.619 "name": "BaseBdev2", 00:16:36.619 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:36.619 "is_configured": true, 00:16:36.619 "data_offset": 256, 00:16:36.619 "data_size": 7936 00:16:36.619 } 00:16:36.619 ] 00:16:36.619 }' 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:36.619 19:06:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.879 19:06:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:36.879 19:06:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:36.879 19:06:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.879 [2024-12-05 19:06:54.399462] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:36.879 [2024-12-05 19:06:54.403163] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:16:36.879 [2024-12-05 19:06:54.405098] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:36.879 19:06:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:36.879 19:06:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:38.259 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:38.259 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:38.259 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:38.259 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:38.259 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:38.259 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:38.259 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:38.259 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:38.259 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:38.259 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:38.259 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:38.259 "name": "raid_bdev1", 00:16:38.259 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:38.259 "strip_size_kb": 0, 00:16:38.259 "state": "online", 00:16:38.259 "raid_level": "raid1", 00:16:38.259 "superblock": true, 00:16:38.259 "num_base_bdevs": 2, 00:16:38.259 "num_base_bdevs_discovered": 2, 00:16:38.259 "num_base_bdevs_operational": 2, 00:16:38.259 "process": { 00:16:38.259 "type": "rebuild", 00:16:38.259 "target": "spare", 00:16:38.259 "progress": { 00:16:38.259 "blocks": 2560, 00:16:38.259 "percent": 32 00:16:38.259 } 00:16:38.259 }, 00:16:38.259 "base_bdevs_list": [ 00:16:38.259 { 00:16:38.259 "name": "spare", 00:16:38.259 "uuid": "39deddf9-8ea1-5e07-9cbe-2346fea21cda", 00:16:38.259 "is_configured": true, 00:16:38.259 "data_offset": 256, 00:16:38.259 "data_size": 7936 00:16:38.259 }, 00:16:38.259 { 00:16:38.259 "name": "BaseBdev2", 00:16:38.259 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:38.259 "is_configured": true, 00:16:38.259 "data_offset": 256, 00:16:38.259 "data_size": 7936 00:16:38.259 } 00:16:38.259 ] 00:16:38.259 }' 00:16:38.259 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:38.259 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:38.260 [2024-12-05 19:06:55.560158] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:38.260 [2024-12-05 19:06:55.609790] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:38.260 [2024-12-05 19:06:55.609843] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:38.260 [2024-12-05 19:06:55.609858] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:38.260 [2024-12-05 19:06:55.609865] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:38.260 "name": "raid_bdev1", 00:16:38.260 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:38.260 "strip_size_kb": 0, 00:16:38.260 "state": "online", 00:16:38.260 "raid_level": "raid1", 00:16:38.260 "superblock": true, 00:16:38.260 "num_base_bdevs": 2, 00:16:38.260 "num_base_bdevs_discovered": 1, 00:16:38.260 "num_base_bdevs_operational": 1, 00:16:38.260 "base_bdevs_list": [ 00:16:38.260 { 00:16:38.260 "name": null, 00:16:38.260 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:38.260 "is_configured": false, 00:16:38.260 "data_offset": 0, 00:16:38.260 "data_size": 7936 00:16:38.260 }, 00:16:38.260 { 00:16:38.260 "name": "BaseBdev2", 00:16:38.260 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:38.260 "is_configured": true, 00:16:38.260 "data_offset": 256, 00:16:38.260 "data_size": 7936 00:16:38.260 } 00:16:38.260 ] 00:16:38.260 }' 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:38.260 19:06:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:38.520 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:38.520 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:38.520 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:38.520 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:38.520 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:38.520 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:38.520 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:38.520 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:38.520 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:38.779 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:38.779 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:38.779 "name": "raid_bdev1", 00:16:38.779 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:38.779 "strip_size_kb": 0, 00:16:38.779 "state": "online", 00:16:38.779 "raid_level": "raid1", 00:16:38.779 "superblock": true, 00:16:38.779 "num_base_bdevs": 2, 00:16:38.779 "num_base_bdevs_discovered": 1, 00:16:38.779 "num_base_bdevs_operational": 1, 00:16:38.779 "base_bdevs_list": [ 00:16:38.779 { 00:16:38.779 "name": null, 00:16:38.779 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:38.779 "is_configured": false, 00:16:38.779 "data_offset": 0, 00:16:38.779 "data_size": 7936 00:16:38.779 }, 00:16:38.779 { 00:16:38.779 "name": "BaseBdev2", 00:16:38.779 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:38.779 "is_configured": true, 00:16:38.779 "data_offset": 256, 00:16:38.779 "data_size": 7936 00:16:38.779 } 00:16:38.779 ] 00:16:38.779 }' 00:16:38.779 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:38.779 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:38.779 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:38.779 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:38.779 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:38.779 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:38.779 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:38.779 [2024-12-05 19:06:56.216837] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:38.779 [2024-12-05 19:06:56.220474] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:38.779 [2024-12-05 19:06:56.222354] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:38.779 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:38.779 19:06:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:39.719 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:39.719 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:39.719 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:39.719 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:39.719 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:39.719 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.719 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:39.719 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.719 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:39.719 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.980 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:39.980 "name": "raid_bdev1", 00:16:39.980 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:39.980 "strip_size_kb": 0, 00:16:39.980 "state": "online", 00:16:39.980 "raid_level": "raid1", 00:16:39.980 "superblock": true, 00:16:39.980 "num_base_bdevs": 2, 00:16:39.980 "num_base_bdevs_discovered": 2, 00:16:39.980 "num_base_bdevs_operational": 2, 00:16:39.980 "process": { 00:16:39.980 "type": "rebuild", 00:16:39.980 "target": "spare", 00:16:39.980 "progress": { 00:16:39.980 "blocks": 2560, 00:16:39.980 "percent": 32 00:16:39.980 } 00:16:39.980 }, 00:16:39.980 "base_bdevs_list": [ 00:16:39.980 { 00:16:39.980 "name": "spare", 00:16:39.980 "uuid": "39deddf9-8ea1-5e07-9cbe-2346fea21cda", 00:16:39.980 "is_configured": true, 00:16:39.980 "data_offset": 256, 00:16:39.980 "data_size": 7936 00:16:39.980 }, 00:16:39.980 { 00:16:39.980 "name": "BaseBdev2", 00:16:39.980 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:39.980 "is_configured": true, 00:16:39.980 "data_offset": 256, 00:16:39.980 "data_size": 7936 00:16:39.980 } 00:16:39.980 ] 00:16:39.980 }' 00:16:39.980 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:39.980 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:39.980 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:39.980 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:39.980 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:16:39.980 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:16:39.980 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:16:39.980 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:16:39.980 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:16:39.980 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:16:39.980 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@706 -- # local timeout=607 00:16:39.980 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:39.980 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:39.980 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:39.980 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:39.981 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:39.981 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:39.981 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.981 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:39.981 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:39.981 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:39.981 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:39.981 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:39.981 "name": "raid_bdev1", 00:16:39.981 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:39.981 "strip_size_kb": 0, 00:16:39.981 "state": "online", 00:16:39.981 "raid_level": "raid1", 00:16:39.981 "superblock": true, 00:16:39.981 "num_base_bdevs": 2, 00:16:39.981 "num_base_bdevs_discovered": 2, 00:16:39.981 "num_base_bdevs_operational": 2, 00:16:39.981 "process": { 00:16:39.981 "type": "rebuild", 00:16:39.981 "target": "spare", 00:16:39.981 "progress": { 00:16:39.981 "blocks": 2816, 00:16:39.981 "percent": 35 00:16:39.981 } 00:16:39.981 }, 00:16:39.981 "base_bdevs_list": [ 00:16:39.981 { 00:16:39.981 "name": "spare", 00:16:39.981 "uuid": "39deddf9-8ea1-5e07-9cbe-2346fea21cda", 00:16:39.981 "is_configured": true, 00:16:39.981 "data_offset": 256, 00:16:39.981 "data_size": 7936 00:16:39.981 }, 00:16:39.981 { 00:16:39.981 "name": "BaseBdev2", 00:16:39.981 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:39.981 "is_configured": true, 00:16:39.981 "data_offset": 256, 00:16:39.981 "data_size": 7936 00:16:39.981 } 00:16:39.981 ] 00:16:39.981 }' 00:16:39.981 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:39.981 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:39.981 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:39.981 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:39.981 19:06:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:41.366 "name": "raid_bdev1", 00:16:41.366 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:41.366 "strip_size_kb": 0, 00:16:41.366 "state": "online", 00:16:41.366 "raid_level": "raid1", 00:16:41.366 "superblock": true, 00:16:41.366 "num_base_bdevs": 2, 00:16:41.366 "num_base_bdevs_discovered": 2, 00:16:41.366 "num_base_bdevs_operational": 2, 00:16:41.366 "process": { 00:16:41.366 "type": "rebuild", 00:16:41.366 "target": "spare", 00:16:41.366 "progress": { 00:16:41.366 "blocks": 5888, 00:16:41.366 "percent": 74 00:16:41.366 } 00:16:41.366 }, 00:16:41.366 "base_bdevs_list": [ 00:16:41.366 { 00:16:41.366 "name": "spare", 00:16:41.366 "uuid": "39deddf9-8ea1-5e07-9cbe-2346fea21cda", 00:16:41.366 "is_configured": true, 00:16:41.366 "data_offset": 256, 00:16:41.366 "data_size": 7936 00:16:41.366 }, 00:16:41.366 { 00:16:41.366 "name": "BaseBdev2", 00:16:41.366 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:41.366 "is_configured": true, 00:16:41.366 "data_offset": 256, 00:16:41.366 "data_size": 7936 00:16:41.366 } 00:16:41.366 ] 00:16:41.366 }' 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:41.366 19:06:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:41.937 [2024-12-05 19:06:59.332799] bdev_raid.c:2900:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:41.937 [2024-12-05 19:06:59.332927] bdev_raid.c:2562:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:41.937 [2024-12-05 19:06:59.333079] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:42.197 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:42.197 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:42.197 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:42.197 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:42.197 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:42.197 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:42.197 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:42.197 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:42.197 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:42.197 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:42.197 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:42.197 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:42.197 "name": "raid_bdev1", 00:16:42.197 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:42.197 "strip_size_kb": 0, 00:16:42.197 "state": "online", 00:16:42.197 "raid_level": "raid1", 00:16:42.197 "superblock": true, 00:16:42.197 "num_base_bdevs": 2, 00:16:42.197 "num_base_bdevs_discovered": 2, 00:16:42.197 "num_base_bdevs_operational": 2, 00:16:42.197 "base_bdevs_list": [ 00:16:42.197 { 00:16:42.197 "name": "spare", 00:16:42.197 "uuid": "39deddf9-8ea1-5e07-9cbe-2346fea21cda", 00:16:42.197 "is_configured": true, 00:16:42.197 "data_offset": 256, 00:16:42.197 "data_size": 7936 00:16:42.197 }, 00:16:42.197 { 00:16:42.197 "name": "BaseBdev2", 00:16:42.197 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:42.197 "is_configured": true, 00:16:42.197 "data_offset": 256, 00:16:42.197 "data_size": 7936 00:16:42.197 } 00:16:42.197 ] 00:16:42.197 }' 00:16:42.197 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@709 -- # break 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:42.458 "name": "raid_bdev1", 00:16:42.458 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:42.458 "strip_size_kb": 0, 00:16:42.458 "state": "online", 00:16:42.458 "raid_level": "raid1", 00:16:42.458 "superblock": true, 00:16:42.458 "num_base_bdevs": 2, 00:16:42.458 "num_base_bdevs_discovered": 2, 00:16:42.458 "num_base_bdevs_operational": 2, 00:16:42.458 "base_bdevs_list": [ 00:16:42.458 { 00:16:42.458 "name": "spare", 00:16:42.458 "uuid": "39deddf9-8ea1-5e07-9cbe-2346fea21cda", 00:16:42.458 "is_configured": true, 00:16:42.458 "data_offset": 256, 00:16:42.458 "data_size": 7936 00:16:42.458 }, 00:16:42.458 { 00:16:42.458 "name": "BaseBdev2", 00:16:42.458 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:42.458 "is_configured": true, 00:16:42.458 "data_offset": 256, 00:16:42.458 "data_size": 7936 00:16:42.458 } 00:16:42.458 ] 00:16:42.458 }' 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:42.458 19:06:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:42.458 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:42.458 "name": "raid_bdev1", 00:16:42.458 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:42.458 "strip_size_kb": 0, 00:16:42.458 "state": "online", 00:16:42.458 "raid_level": "raid1", 00:16:42.458 "superblock": true, 00:16:42.458 "num_base_bdevs": 2, 00:16:42.458 "num_base_bdevs_discovered": 2, 00:16:42.458 "num_base_bdevs_operational": 2, 00:16:42.458 "base_bdevs_list": [ 00:16:42.458 { 00:16:42.458 "name": "spare", 00:16:42.458 "uuid": "39deddf9-8ea1-5e07-9cbe-2346fea21cda", 00:16:42.458 "is_configured": true, 00:16:42.458 "data_offset": 256, 00:16:42.458 "data_size": 7936 00:16:42.458 }, 00:16:42.458 { 00:16:42.458 "name": "BaseBdev2", 00:16:42.458 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:42.458 "is_configured": true, 00:16:42.458 "data_offset": 256, 00:16:42.458 "data_size": 7936 00:16:42.458 } 00:16:42.458 ] 00:16:42.458 }' 00:16:42.458 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:42.458 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.029 [2024-12-05 19:07:00.407331] bdev_raid.c:2411:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:43.029 [2024-12-05 19:07:00.407360] bdev_raid.c:1899:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:43.029 [2024-12-05 19:07:00.407457] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:43.029 [2024-12-05 19:07:00.407517] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:43.029 [2024-12-05 19:07:00.407528] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # jq length 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.029 [2024-12-05 19:07:00.483200] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:43.029 [2024-12-05 19:07:00.483253] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:43.029 [2024-12-05 19:07:00.483271] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:16:43.029 [2024-12-05 19:07:00.483282] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:43.029 [2024-12-05 19:07:00.485192] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:43.029 [2024-12-05 19:07:00.485286] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:43.029 [2024-12-05 19:07:00.485343] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:43.029 [2024-12-05 19:07:00.485398] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:43.029 [2024-12-05 19:07:00.485494] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:43.029 spare 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:43.029 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.030 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.030 [2024-12-05 19:07:00.585380] bdev_raid.c:1734:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:16:43.030 [2024-12-05 19:07:00.585404] bdev_raid.c:1735:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:43.030 [2024-12-05 19:07:00.585490] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:16:43.030 [2024-12-05 19:07:00.585560] bdev_raid.c:1764:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:16:43.030 [2024-12-05 19:07:00.585571] bdev_raid.c:1765:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:16:43.030 [2024-12-05 19:07:00.585637] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:43.290 "name": "raid_bdev1", 00:16:43.290 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:43.290 "strip_size_kb": 0, 00:16:43.290 "state": "online", 00:16:43.290 "raid_level": "raid1", 00:16:43.290 "superblock": true, 00:16:43.290 "num_base_bdevs": 2, 00:16:43.290 "num_base_bdevs_discovered": 2, 00:16:43.290 "num_base_bdevs_operational": 2, 00:16:43.290 "base_bdevs_list": [ 00:16:43.290 { 00:16:43.290 "name": "spare", 00:16:43.290 "uuid": "39deddf9-8ea1-5e07-9cbe-2346fea21cda", 00:16:43.290 "is_configured": true, 00:16:43.290 "data_offset": 256, 00:16:43.290 "data_size": 7936 00:16:43.290 }, 00:16:43.290 { 00:16:43.290 "name": "BaseBdev2", 00:16:43.290 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:43.290 "is_configured": true, 00:16:43.290 "data_offset": 256, 00:16:43.290 "data_size": 7936 00:16:43.290 } 00:16:43.290 ] 00:16:43.290 }' 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:43.290 19:07:00 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.550 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:43.550 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:43.550 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:43.550 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:43.550 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:43.550 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:43.550 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:43.550 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.550 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.550 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.550 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:43.550 "name": "raid_bdev1", 00:16:43.550 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:43.550 "strip_size_kb": 0, 00:16:43.550 "state": "online", 00:16:43.550 "raid_level": "raid1", 00:16:43.550 "superblock": true, 00:16:43.550 "num_base_bdevs": 2, 00:16:43.550 "num_base_bdevs_discovered": 2, 00:16:43.550 "num_base_bdevs_operational": 2, 00:16:43.550 "base_bdevs_list": [ 00:16:43.550 { 00:16:43.550 "name": "spare", 00:16:43.550 "uuid": "39deddf9-8ea1-5e07-9cbe-2346fea21cda", 00:16:43.550 "is_configured": true, 00:16:43.550 "data_offset": 256, 00:16:43.550 "data_size": 7936 00:16:43.550 }, 00:16:43.550 { 00:16:43.550 "name": "BaseBdev2", 00:16:43.550 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:43.550 "is_configured": true, 00:16:43.550 "data_offset": 256, 00:16:43.550 "data_size": 7936 00:16:43.550 } 00:16:43.550 ] 00:16:43.550 }' 00:16:43.550 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.809 [2024-12-05 19:07:01.261970] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:43.809 "name": "raid_bdev1", 00:16:43.809 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:43.809 "strip_size_kb": 0, 00:16:43.809 "state": "online", 00:16:43.809 "raid_level": "raid1", 00:16:43.809 "superblock": true, 00:16:43.809 "num_base_bdevs": 2, 00:16:43.809 "num_base_bdevs_discovered": 1, 00:16:43.809 "num_base_bdevs_operational": 1, 00:16:43.809 "base_bdevs_list": [ 00:16:43.809 { 00:16:43.809 "name": null, 00:16:43.809 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:43.809 "is_configured": false, 00:16:43.809 "data_offset": 0, 00:16:43.809 "data_size": 7936 00:16:43.809 }, 00:16:43.809 { 00:16:43.809 "name": "BaseBdev2", 00:16:43.809 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:43.809 "is_configured": true, 00:16:43.809 "data_offset": 256, 00:16:43.809 "data_size": 7936 00:16:43.809 } 00:16:43.809 ] 00:16:43.809 }' 00:16:43.809 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:43.810 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:44.377 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:44.377 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:44.377 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:44.377 [2024-12-05 19:07:01.733207] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:44.377 [2024-12-05 19:07:01.733356] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:44.377 [2024-12-05 19:07:01.733371] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:44.377 [2024-12-05 19:07:01.733406] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:44.377 [2024-12-05 19:07:01.736916] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:16:44.377 [2024-12-05 19:07:01.738822] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:44.377 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:44.377 19:07:01 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:45.318 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:45.318 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:45.318 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:45.318 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:45.318 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:45.318 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:45.318 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:45.318 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:45.318 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:45.318 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:45.318 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:45.318 "name": "raid_bdev1", 00:16:45.318 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:45.318 "strip_size_kb": 0, 00:16:45.318 "state": "online", 00:16:45.318 "raid_level": "raid1", 00:16:45.318 "superblock": true, 00:16:45.318 "num_base_bdevs": 2, 00:16:45.318 "num_base_bdevs_discovered": 2, 00:16:45.318 "num_base_bdevs_operational": 2, 00:16:45.318 "process": { 00:16:45.318 "type": "rebuild", 00:16:45.318 "target": "spare", 00:16:45.318 "progress": { 00:16:45.318 "blocks": 2560, 00:16:45.318 "percent": 32 00:16:45.318 } 00:16:45.318 }, 00:16:45.318 "base_bdevs_list": [ 00:16:45.318 { 00:16:45.318 "name": "spare", 00:16:45.318 "uuid": "39deddf9-8ea1-5e07-9cbe-2346fea21cda", 00:16:45.318 "is_configured": true, 00:16:45.318 "data_offset": 256, 00:16:45.318 "data_size": 7936 00:16:45.318 }, 00:16:45.318 { 00:16:45.318 "name": "BaseBdev2", 00:16:45.318 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:45.318 "is_configured": true, 00:16:45.318 "data_offset": 256, 00:16:45.318 "data_size": 7936 00:16:45.318 } 00:16:45.318 ] 00:16:45.318 }' 00:16:45.318 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:45.318 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:45.318 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:45.578 [2024-12-05 19:07:02.901571] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:45.578 [2024-12-05 19:07:02.942756] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:45.578 [2024-12-05 19:07:02.942807] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:45.578 [2024-12-05 19:07:02.942823] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:45.578 [2024-12-05 19:07:02.942829] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:45.578 19:07:02 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:45.578 19:07:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:45.578 "name": "raid_bdev1", 00:16:45.578 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:45.578 "strip_size_kb": 0, 00:16:45.578 "state": "online", 00:16:45.578 "raid_level": "raid1", 00:16:45.578 "superblock": true, 00:16:45.578 "num_base_bdevs": 2, 00:16:45.578 "num_base_bdevs_discovered": 1, 00:16:45.578 "num_base_bdevs_operational": 1, 00:16:45.578 "base_bdevs_list": [ 00:16:45.578 { 00:16:45.578 "name": null, 00:16:45.578 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:45.578 "is_configured": false, 00:16:45.578 "data_offset": 0, 00:16:45.578 "data_size": 7936 00:16:45.578 }, 00:16:45.578 { 00:16:45.578 "name": "BaseBdev2", 00:16:45.578 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:45.578 "is_configured": true, 00:16:45.578 "data_offset": 256, 00:16:45.578 "data_size": 7936 00:16:45.578 } 00:16:45.578 ] 00:16:45.578 }' 00:16:45.578 19:07:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:45.578 19:07:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:46.148 19:07:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:46.148 19:07:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:46.148 19:07:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:46.148 [2024-12-05 19:07:03.401867] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:46.148 [2024-12-05 19:07:03.401963] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:46.148 [2024-12-05 19:07:03.402020] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:16:46.148 [2024-12-05 19:07:03.402058] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:46.148 [2024-12-05 19:07:03.402255] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:46.148 [2024-12-05 19:07:03.402301] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:46.148 [2024-12-05 19:07:03.402373] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:46.148 [2024-12-05 19:07:03.402406] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:46.148 [2024-12-05 19:07:03.402445] bdev_raid.c:3758:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:46.148 [2024-12-05 19:07:03.402542] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:46.148 [2024-12-05 19:07:03.405287] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:16:46.148 [2024-12-05 19:07:03.407186] bdev_raid.c:2935:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:46.148 spare 00:16:46.148 19:07:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:46.148 19:07:03 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:47.089 "name": "raid_bdev1", 00:16:47.089 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:47.089 "strip_size_kb": 0, 00:16:47.089 "state": "online", 00:16:47.089 "raid_level": "raid1", 00:16:47.089 "superblock": true, 00:16:47.089 "num_base_bdevs": 2, 00:16:47.089 "num_base_bdevs_discovered": 2, 00:16:47.089 "num_base_bdevs_operational": 2, 00:16:47.089 "process": { 00:16:47.089 "type": "rebuild", 00:16:47.089 "target": "spare", 00:16:47.089 "progress": { 00:16:47.089 "blocks": 2560, 00:16:47.089 "percent": 32 00:16:47.089 } 00:16:47.089 }, 00:16:47.089 "base_bdevs_list": [ 00:16:47.089 { 00:16:47.089 "name": "spare", 00:16:47.089 "uuid": "39deddf9-8ea1-5e07-9cbe-2346fea21cda", 00:16:47.089 "is_configured": true, 00:16:47.089 "data_offset": 256, 00:16:47.089 "data_size": 7936 00:16:47.089 }, 00:16:47.089 { 00:16:47.089 "name": "BaseBdev2", 00:16:47.089 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:47.089 "is_configured": true, 00:16:47.089 "data_offset": 256, 00:16:47.089 "data_size": 7936 00:16:47.089 } 00:16:47.089 ] 00:16:47.089 }' 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:47.089 [2024-12-05 19:07:04.554366] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:47.089 [2024-12-05 19:07:04.611174] bdev_raid.c:2571:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:47.089 [2024-12-05 19:07:04.611298] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:47.089 [2024-12-05 19:07:04.611314] bdev_raid.c:2175:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:47.089 [2024-12-05 19:07:04.611324] bdev_raid.c:2509:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:47.089 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.368 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:47.368 "name": "raid_bdev1", 00:16:47.368 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:47.368 "strip_size_kb": 0, 00:16:47.368 "state": "online", 00:16:47.368 "raid_level": "raid1", 00:16:47.368 "superblock": true, 00:16:47.368 "num_base_bdevs": 2, 00:16:47.368 "num_base_bdevs_discovered": 1, 00:16:47.368 "num_base_bdevs_operational": 1, 00:16:47.368 "base_bdevs_list": [ 00:16:47.368 { 00:16:47.368 "name": null, 00:16:47.368 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:47.368 "is_configured": false, 00:16:47.368 "data_offset": 0, 00:16:47.368 "data_size": 7936 00:16:47.368 }, 00:16:47.368 { 00:16:47.368 "name": "BaseBdev2", 00:16:47.368 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:47.368 "is_configured": true, 00:16:47.368 "data_offset": 256, 00:16:47.368 "data_size": 7936 00:16:47.368 } 00:16:47.368 ] 00:16:47.368 }' 00:16:47.368 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:47.368 19:07:04 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:47.627 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:47.627 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:47.627 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:47.627 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:47.627 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:47.627 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:47.627 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:47.627 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.627 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:47.627 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.627 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:47.627 "name": "raid_bdev1", 00:16:47.627 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:47.627 "strip_size_kb": 0, 00:16:47.627 "state": "online", 00:16:47.627 "raid_level": "raid1", 00:16:47.627 "superblock": true, 00:16:47.627 "num_base_bdevs": 2, 00:16:47.627 "num_base_bdevs_discovered": 1, 00:16:47.627 "num_base_bdevs_operational": 1, 00:16:47.627 "base_bdevs_list": [ 00:16:47.627 { 00:16:47.627 "name": null, 00:16:47.627 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:47.627 "is_configured": false, 00:16:47.627 "data_offset": 0, 00:16:47.627 "data_size": 7936 00:16:47.627 }, 00:16:47.627 { 00:16:47.627 "name": "BaseBdev2", 00:16:47.627 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:47.627 "is_configured": true, 00:16:47.627 "data_offset": 256, 00:16:47.627 "data_size": 7936 00:16:47.627 } 00:16:47.628 ] 00:16:47.628 }' 00:16:47.628 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:47.628 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:47.628 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:47.628 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:47.628 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:16:47.628 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.628 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:47.887 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.887 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:47.887 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:47.887 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:47.887 [2024-12-05 19:07:05.197941] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:47.887 [2024-12-05 19:07:05.197996] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:47.887 [2024-12-05 19:07:05.198016] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:16:47.887 [2024-12-05 19:07:05.198027] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:47.887 [2024-12-05 19:07:05.198181] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:47.887 [2024-12-05 19:07:05.198196] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:47.887 [2024-12-05 19:07:05.198239] bdev_raid.c:3907:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:16:47.887 [2024-12-05 19:07:05.198252] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:47.887 [2024-12-05 19:07:05.198260] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:47.887 [2024-12-05 19:07:05.198273] bdev_raid.c:3894:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:16:47.887 BaseBdev1 00:16:47.887 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:47.887 19:07:05 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@775 -- # sleep 1 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:48.825 "name": "raid_bdev1", 00:16:48.825 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:48.825 "strip_size_kb": 0, 00:16:48.825 "state": "online", 00:16:48.825 "raid_level": "raid1", 00:16:48.825 "superblock": true, 00:16:48.825 "num_base_bdevs": 2, 00:16:48.825 "num_base_bdevs_discovered": 1, 00:16:48.825 "num_base_bdevs_operational": 1, 00:16:48.825 "base_bdevs_list": [ 00:16:48.825 { 00:16:48.825 "name": null, 00:16:48.825 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:48.825 "is_configured": false, 00:16:48.825 "data_offset": 0, 00:16:48.825 "data_size": 7936 00:16:48.825 }, 00:16:48.825 { 00:16:48.825 "name": "BaseBdev2", 00:16:48.825 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:48.825 "is_configured": true, 00:16:48.825 "data_offset": 256, 00:16:48.825 "data_size": 7936 00:16:48.825 } 00:16:48.825 ] 00:16:48.825 }' 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:48.825 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:49.085 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:49.085 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:49.345 "name": "raid_bdev1", 00:16:49.345 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:49.345 "strip_size_kb": 0, 00:16:49.345 "state": "online", 00:16:49.345 "raid_level": "raid1", 00:16:49.345 "superblock": true, 00:16:49.345 "num_base_bdevs": 2, 00:16:49.345 "num_base_bdevs_discovered": 1, 00:16:49.345 "num_base_bdevs_operational": 1, 00:16:49.345 "base_bdevs_list": [ 00:16:49.345 { 00:16:49.345 "name": null, 00:16:49.345 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:49.345 "is_configured": false, 00:16:49.345 "data_offset": 0, 00:16:49.345 "data_size": 7936 00:16:49.345 }, 00:16:49.345 { 00:16:49.345 "name": "BaseBdev2", 00:16:49.345 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:49.345 "is_configured": true, 00:16:49.345 "data_offset": 256, 00:16:49.345 "data_size": 7936 00:16:49.345 } 00:16:49.345 ] 00:16:49.345 }' 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@652 -- # local es=0 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@655 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:49.345 [2024-12-05 19:07:06.807226] bdev_raid.c:3326:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:49.345 [2024-12-05 19:07:06.807439] bdev_raid.c:3700:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:49.345 [2024-12-05 19:07:06.807464] bdev_raid.c:3719:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:49.345 request: 00:16:49.345 { 00:16:49.345 "base_bdev": "BaseBdev1", 00:16:49.345 "raid_bdev": "raid_bdev1", 00:16:49.345 "method": "bdev_raid_add_base_bdev", 00:16:49.345 "req_id": 1 00:16:49.345 } 00:16:49.345 Got JSON-RPC error response 00:16:49.345 response: 00:16:49.345 { 00:16:49.345 "code": -22, 00:16:49.345 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:49.345 } 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@655 -- # es=1 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:16:49.345 19:07:06 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:50.286 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:50.286 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:50.286 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:50.286 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:50.286 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:50.286 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:50.286 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:50.286 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:50.286 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:50.286 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:50.286 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:50.286 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:50.286 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:50.286 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:50.545 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:50.545 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:50.545 "name": "raid_bdev1", 00:16:50.545 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:50.545 "strip_size_kb": 0, 00:16:50.545 "state": "online", 00:16:50.545 "raid_level": "raid1", 00:16:50.545 "superblock": true, 00:16:50.545 "num_base_bdevs": 2, 00:16:50.545 "num_base_bdevs_discovered": 1, 00:16:50.545 "num_base_bdevs_operational": 1, 00:16:50.545 "base_bdevs_list": [ 00:16:50.545 { 00:16:50.545 "name": null, 00:16:50.545 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:50.545 "is_configured": false, 00:16:50.545 "data_offset": 0, 00:16:50.545 "data_size": 7936 00:16:50.545 }, 00:16:50.545 { 00:16:50.545 "name": "BaseBdev2", 00:16:50.545 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:50.545 "is_configured": true, 00:16:50.546 "data_offset": 256, 00:16:50.546 "data_size": 7936 00:16:50.546 } 00:16:50.546 ] 00:16:50.546 }' 00:16:50.546 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:50.546 19:07:07 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:50.804 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:50.804 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:50.804 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:50.804 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:50.804 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:50.804 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:50.804 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:50.804 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@563 -- # xtrace_disable 00:16:50.804 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:50.804 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:16:50.804 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:50.804 "name": "raid_bdev1", 00:16:50.804 "uuid": "3609e837-91db-482e-9c73-64395b95bcbc", 00:16:50.804 "strip_size_kb": 0, 00:16:50.804 "state": "online", 00:16:50.804 "raid_level": "raid1", 00:16:50.804 "superblock": true, 00:16:50.804 "num_base_bdevs": 2, 00:16:50.804 "num_base_bdevs_discovered": 1, 00:16:50.804 "num_base_bdevs_operational": 1, 00:16:50.804 "base_bdevs_list": [ 00:16:50.804 { 00:16:50.804 "name": null, 00:16:50.804 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:50.804 "is_configured": false, 00:16:50.804 "data_offset": 0, 00:16:50.804 "data_size": 7936 00:16:50.804 }, 00:16:50.804 { 00:16:50.804 "name": "BaseBdev2", 00:16:50.804 "uuid": "09e57450-c233-52cc-b680-cb1ddfd3dfb4", 00:16:50.804 "is_configured": true, 00:16:50.804 "data_offset": 256, 00:16:50.804 "data_size": 7936 00:16:50.804 } 00:16:50.804 ] 00:16:50.804 }' 00:16:50.804 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:50.804 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:50.804 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:51.063 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:51.063 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@784 -- # killprocess 96137 00:16:51.063 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # '[' -z 96137 ']' 00:16:51.063 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@958 -- # kill -0 96137 00:16:51.064 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # uname 00:16:51.064 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:51.064 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 96137 00:16:51.064 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:51.064 killing process with pid 96137 00:16:51.064 Received shutdown signal, test time was about 60.000000 seconds 00:16:51.064 00:16:51.064 Latency(us) 00:16:51.064 [2024-12-05T19:07:08.623Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:51.064 [2024-12-05T19:07:08.623Z] =================================================================================================================== 00:16:51.064 [2024-12-05T19:07:08.623Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:51.064 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:51.064 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@972 -- # echo 'killing process with pid 96137' 00:16:51.064 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@973 -- # kill 96137 00:16:51.064 [2024-12-05 19:07:08.419833] bdev_raid.c:1387:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:51.064 [2024-12-05 19:07:08.419938] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:51.064 [2024-12-05 19:07:08.419987] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:51.064 [2024-12-05 19:07:08.419996] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:16:51.064 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@978 -- # wait 96137 00:16:51.064 [2024-12-05 19:07:08.453096] bdev_raid.c:1413:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:51.323 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@786 -- # return 0 00:16:51.323 00:16:51.323 real 0m16.275s 00:16:51.323 user 0m21.843s 00:16:51.323 sys 0m1.673s 00:16:51.323 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:51.323 ************************************ 00:16:51.323 END TEST raid_rebuild_test_sb_md_interleaved 00:16:51.323 ************************************ 00:16:51.323 19:07:08 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:51.323 19:07:08 bdev_raid -- bdev/bdev_raid.sh@1015 -- # trap - EXIT 00:16:51.323 19:07:08 bdev_raid -- bdev/bdev_raid.sh@1016 -- # cleanup 00:16:51.323 19:07:08 bdev_raid -- bdev/bdev_raid.sh@56 -- # '[' -n 96137 ']' 00:16:51.323 19:07:08 bdev_raid -- bdev/bdev_raid.sh@56 -- # ps -p 96137 00:16:51.323 19:07:08 bdev_raid -- bdev/bdev_raid.sh@60 -- # rm -rf /raidtest 00:16:51.323 00:16:51.323 real 9m47.939s 00:16:51.323 user 13m57.903s 00:16:51.323 sys 1m50.632s 00:16:51.323 ************************************ 00:16:51.323 END TEST bdev_raid 00:16:51.323 ************************************ 00:16:51.323 19:07:08 bdev_raid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:51.323 19:07:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:51.323 19:07:08 -- spdk/autotest.sh@190 -- # run_test spdkcli_raid /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:16:51.323 19:07:08 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:16:51.323 19:07:08 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:51.323 19:07:08 -- common/autotest_common.sh@10 -- # set +x 00:16:51.323 ************************************ 00:16:51.323 START TEST spdkcli_raid 00:16:51.323 ************************************ 00:16:51.323 19:07:08 spdkcli_raid -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:16:51.583 * Looking for test storage... 00:16:51.583 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:16:51.583 19:07:08 spdkcli_raid -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:16:51.584 19:07:08 spdkcli_raid -- common/autotest_common.sh@1711 -- # lcov --version 00:16:51.584 19:07:08 spdkcli_raid -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:16:51.584 19:07:09 spdkcli_raid -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@336 -- # IFS=.-: 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@336 -- # read -ra ver1 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@337 -- # IFS=.-: 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@337 -- # read -ra ver2 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@338 -- # local 'op=<' 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@340 -- # ver1_l=2 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@341 -- # ver2_l=1 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@344 -- # case "$op" in 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@345 -- # : 1 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@365 -- # decimal 1 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@353 -- # local d=1 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@355 -- # echo 1 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@366 -- # decimal 2 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@353 -- # local d=2 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@355 -- # echo 2 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:51.584 19:07:09 spdkcli_raid -- scripts/common.sh@368 -- # return 0 00:16:51.584 19:07:09 spdkcli_raid -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:51.584 19:07:09 spdkcli_raid -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:16:51.584 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:51.584 --rc genhtml_branch_coverage=1 00:16:51.584 --rc genhtml_function_coverage=1 00:16:51.584 --rc genhtml_legend=1 00:16:51.584 --rc geninfo_all_blocks=1 00:16:51.584 --rc geninfo_unexecuted_blocks=1 00:16:51.584 00:16:51.584 ' 00:16:51.584 19:07:09 spdkcli_raid -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:16:51.584 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:51.584 --rc genhtml_branch_coverage=1 00:16:51.584 --rc genhtml_function_coverage=1 00:16:51.584 --rc genhtml_legend=1 00:16:51.584 --rc geninfo_all_blocks=1 00:16:51.584 --rc geninfo_unexecuted_blocks=1 00:16:51.584 00:16:51.584 ' 00:16:51.584 19:07:09 spdkcli_raid -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:16:51.584 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:51.584 --rc genhtml_branch_coverage=1 00:16:51.584 --rc genhtml_function_coverage=1 00:16:51.584 --rc genhtml_legend=1 00:16:51.584 --rc geninfo_all_blocks=1 00:16:51.584 --rc geninfo_unexecuted_blocks=1 00:16:51.584 00:16:51.584 ' 00:16:51.584 19:07:09 spdkcli_raid -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:16:51.584 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:51.584 --rc genhtml_branch_coverage=1 00:16:51.584 --rc genhtml_function_coverage=1 00:16:51.584 --rc genhtml_legend=1 00:16:51.584 --rc geninfo_all_blocks=1 00:16:51.584 --rc geninfo_unexecuted_blocks=1 00:16:51.584 00:16:51.584 ' 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/raid.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/raid.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/iscsi_tgt/common.sh 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@9 -- # ISCSI_BRIDGE=iscsi_br 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@10 -- # INITIATOR_INTERFACE=spdk_init_int 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@11 -- # INITIATOR_BRIDGE=init_br 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@12 -- # TARGET_NAMESPACE=spdk_iscsi_ns 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@13 -- # TARGET_NS_CMD=(ip netns exec "$TARGET_NAMESPACE") 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@14 -- # TARGET_INTERFACE=spdk_tgt_int 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@15 -- # TARGET_INTERFACE2=spdk_tgt_int2 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@16 -- # TARGET_BRIDGE=tgt_br 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@17 -- # TARGET_BRIDGE2=tgt_br2 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@20 -- # TARGET_IP=10.0.0.1 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@21 -- # TARGET_IP2=10.0.0.3 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@22 -- # INITIATOR_IP=10.0.0.2 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@23 -- # ISCSI_PORT=3260 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@24 -- # NETMASK=10.0.0.2/32 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@25 -- # INITIATOR_TAG=2 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@26 -- # INITIATOR_NAME=ANY 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@27 -- # PORTAL_TAG=1 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@28 -- # ISCSI_APP=("${TARGET_NS_CMD[@]}" "${ISCSI_APP[@]}") 00:16:51.584 19:07:09 spdkcli_raid -- iscsi_tgt/common.sh@29 -- # ISCSI_TEST_CORE_MASK=0xF 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/raid.sh@12 -- # MATCH_FILE=spdkcli_raid.test 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/raid.sh@13 -- # SPDKCLI_BRANCH=/bdevs 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/raid.sh@14 -- # dirname /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/raid.sh@14 -- # readlink -f /home/vagrant/spdk_repo/spdk/test/spdkcli 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/raid.sh@14 -- # testdir=/home/vagrant/spdk_repo/spdk/test/spdkcli 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/raid.sh@15 -- # . /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/raid.sh@17 -- # trap cleanup EXIT 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/raid.sh@19 -- # timing_enter run_spdk_tgt 00:16:51.584 19:07:09 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:51.584 19:07:09 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/raid.sh@20 -- # run_spdk_tgt 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/common.sh@27 -- # spdk_tgt_pid=96709 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/common.sh@26 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:16:51.584 19:07:09 spdkcli_raid -- spdkcli/common.sh@28 -- # waitforlisten 96709 00:16:51.584 19:07:09 spdkcli_raid -- common/autotest_common.sh@835 -- # '[' -z 96709 ']' 00:16:51.584 19:07:09 spdkcli_raid -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:51.584 19:07:09 spdkcli_raid -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:51.584 19:07:09 spdkcli_raid -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:51.584 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:51.584 19:07:09 spdkcli_raid -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:51.584 19:07:09 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:51.845 [2024-12-05 19:07:09.198790] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:16:51.845 [2024-12-05 19:07:09.199007] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96709 ] 00:16:51.845 [2024-12-05 19:07:09.356375] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:16:51.845 [2024-12-05 19:07:09.383794] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:51.845 [2024-12-05 19:07:09.383879] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:16:52.784 19:07:10 spdkcli_raid -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:16:52.785 19:07:10 spdkcli_raid -- common/autotest_common.sh@868 -- # return 0 00:16:52.785 19:07:10 spdkcli_raid -- spdkcli/raid.sh@21 -- # timing_exit run_spdk_tgt 00:16:52.785 19:07:10 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:52.785 19:07:10 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:52.785 19:07:10 spdkcli_raid -- spdkcli/raid.sh@23 -- # timing_enter spdkcli_create_malloc 00:16:52.785 19:07:10 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:52.785 19:07:10 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:52.785 19:07:10 spdkcli_raid -- spdkcli/raid.sh@26 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 8 512 Malloc1'\'' '\''Malloc1'\'' True 00:16:52.785 '\''/bdevs/malloc create 8 512 Malloc2'\'' '\''Malloc2'\'' True 00:16:52.785 ' 00:16:54.166 Executing command: ['/bdevs/malloc create 8 512 Malloc1', 'Malloc1', True] 00:16:54.166 Executing command: ['/bdevs/malloc create 8 512 Malloc2', 'Malloc2', True] 00:16:54.166 19:07:11 spdkcli_raid -- spdkcli/raid.sh@27 -- # timing_exit spdkcli_create_malloc 00:16:54.166 19:07:11 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:54.166 19:07:11 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:54.425 19:07:11 spdkcli_raid -- spdkcli/raid.sh@29 -- # timing_enter spdkcli_create_raid 00:16:54.425 19:07:11 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:54.425 19:07:11 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:54.425 19:07:11 spdkcli_raid -- spdkcli/raid.sh@31 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4'\'' '\''testraid'\'' True 00:16:54.425 ' 00:16:55.364 Executing command: ['/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4', 'testraid', True] 00:16:55.364 19:07:12 spdkcli_raid -- spdkcli/raid.sh@32 -- # timing_exit spdkcli_create_raid 00:16:55.364 19:07:12 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:55.364 19:07:12 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:55.623 19:07:12 spdkcli_raid -- spdkcli/raid.sh@34 -- # timing_enter spdkcli_check_match 00:16:55.623 19:07:12 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:55.623 19:07:12 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:55.623 19:07:12 spdkcli_raid -- spdkcli/raid.sh@35 -- # check_match 00:16:55.623 19:07:12 spdkcli_raid -- spdkcli/common.sh@44 -- # /home/vagrant/spdk_repo/spdk/scripts/spdkcli.py ll /bdevs 00:16:55.882 19:07:13 spdkcli_raid -- spdkcli/common.sh@45 -- # /home/vagrant/spdk_repo/spdk/test/app/match/match /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test.match 00:16:56.141 19:07:13 spdkcli_raid -- spdkcli/common.sh@46 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test 00:16:56.141 19:07:13 spdkcli_raid -- spdkcli/raid.sh@36 -- # timing_exit spdkcli_check_match 00:16:56.141 19:07:13 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:56.141 19:07:13 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:56.141 19:07:13 spdkcli_raid -- spdkcli/raid.sh@38 -- # timing_enter spdkcli_delete_raid 00:16:56.141 19:07:13 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:56.141 19:07:13 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:56.141 19:07:13 spdkcli_raid -- spdkcli/raid.sh@40 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume delete testraid'\'' '\'''\'' True 00:16:56.141 ' 00:16:57.089 Executing command: ['/bdevs/raid_volume delete testraid', '', True] 00:16:57.089 19:07:14 spdkcli_raid -- spdkcli/raid.sh@41 -- # timing_exit spdkcli_delete_raid 00:16:57.089 19:07:14 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:57.089 19:07:14 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:57.369 19:07:14 spdkcli_raid -- spdkcli/raid.sh@43 -- # timing_enter spdkcli_delete_malloc 00:16:57.369 19:07:14 spdkcli_raid -- common/autotest_common.sh@726 -- # xtrace_disable 00:16:57.369 19:07:14 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:57.369 19:07:14 spdkcli_raid -- spdkcli/raid.sh@46 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc delete Malloc1'\'' '\'''\'' True 00:16:57.369 '\''/bdevs/malloc delete Malloc2'\'' '\'''\'' True 00:16:57.369 ' 00:16:58.767 Executing command: ['/bdevs/malloc delete Malloc1', '', True] 00:16:58.767 Executing command: ['/bdevs/malloc delete Malloc2', '', True] 00:16:58.767 19:07:16 spdkcli_raid -- spdkcli/raid.sh@47 -- # timing_exit spdkcli_delete_malloc 00:16:58.767 19:07:16 spdkcli_raid -- common/autotest_common.sh@732 -- # xtrace_disable 00:16:58.767 19:07:16 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:58.767 19:07:16 spdkcli_raid -- spdkcli/raid.sh@49 -- # killprocess 96709 00:16:58.767 19:07:16 spdkcli_raid -- common/autotest_common.sh@954 -- # '[' -z 96709 ']' 00:16:58.767 19:07:16 spdkcli_raid -- common/autotest_common.sh@958 -- # kill -0 96709 00:16:58.767 19:07:16 spdkcli_raid -- common/autotest_common.sh@959 -- # uname 00:16:58.767 19:07:16 spdkcli_raid -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:16:58.767 19:07:16 spdkcli_raid -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 96709 00:16:58.767 19:07:16 spdkcli_raid -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:16:58.767 19:07:16 spdkcli_raid -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:16:58.767 19:07:16 spdkcli_raid -- common/autotest_common.sh@972 -- # echo 'killing process with pid 96709' 00:16:58.767 killing process with pid 96709 00:16:58.767 19:07:16 spdkcli_raid -- common/autotest_common.sh@973 -- # kill 96709 00:16:58.767 19:07:16 spdkcli_raid -- common/autotest_common.sh@978 -- # wait 96709 00:16:59.026 19:07:16 spdkcli_raid -- spdkcli/raid.sh@1 -- # cleanup 00:16:59.026 19:07:16 spdkcli_raid -- spdkcli/common.sh@10 -- # '[' -n 96709 ']' 00:16:59.026 19:07:16 spdkcli_raid -- spdkcli/common.sh@11 -- # killprocess 96709 00:16:59.026 Process with pid 96709 is not found 00:16:59.026 19:07:16 spdkcli_raid -- common/autotest_common.sh@954 -- # '[' -z 96709 ']' 00:16:59.026 19:07:16 spdkcli_raid -- common/autotest_common.sh@958 -- # kill -0 96709 00:16:59.026 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 958: kill: (96709) - No such process 00:16:59.026 19:07:16 spdkcli_raid -- common/autotest_common.sh@981 -- # echo 'Process with pid 96709 is not found' 00:16:59.026 19:07:16 spdkcli_raid -- spdkcli/common.sh@13 -- # '[' -n '' ']' 00:16:59.026 19:07:16 spdkcli_raid -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:16:59.026 19:07:16 spdkcli_raid -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:16:59.026 19:07:16 spdkcli_raid -- spdkcli/common.sh@22 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_raid.test /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:16:59.026 00:16:59.026 real 0m7.757s 00:16:59.026 user 0m16.370s 00:16:59.026 sys 0m1.126s 00:16:59.026 19:07:16 spdkcli_raid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:16:59.026 19:07:16 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:59.026 ************************************ 00:16:59.026 END TEST spdkcli_raid 00:16:59.026 ************************************ 00:16:59.286 19:07:16 -- spdk/autotest.sh@191 -- # run_test blockdev_raid5f /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:16:59.286 19:07:16 -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:16:59.286 19:07:16 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:16:59.286 19:07:16 -- common/autotest_common.sh@10 -- # set +x 00:16:59.286 ************************************ 00:16:59.286 START TEST blockdev_raid5f 00:16:59.286 ************************************ 00:16:59.286 19:07:16 blockdev_raid5f -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:16:59.286 * Looking for test storage... 00:16:59.286 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:16:59.286 19:07:16 blockdev_raid5f -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:16:59.286 19:07:16 blockdev_raid5f -- common/autotest_common.sh@1711 -- # lcov --version 00:16:59.286 19:07:16 blockdev_raid5f -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:16:59.546 19:07:16 blockdev_raid5f -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@336 -- # IFS=.-: 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@336 -- # read -ra ver1 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@337 -- # IFS=.-: 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@337 -- # read -ra ver2 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@338 -- # local 'op=<' 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@340 -- # ver1_l=2 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@341 -- # ver2_l=1 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@344 -- # case "$op" in 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@345 -- # : 1 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@365 -- # decimal 1 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@353 -- # local d=1 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@355 -- # echo 1 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@365 -- # ver1[v]=1 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@366 -- # decimal 2 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@353 -- # local d=2 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@355 -- # echo 2 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@366 -- # ver2[v]=2 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:59.546 19:07:16 blockdev_raid5f -- scripts/common.sh@368 -- # return 0 00:16:59.546 19:07:16 blockdev_raid5f -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:59.546 19:07:16 blockdev_raid5f -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:16:59.546 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:59.546 --rc genhtml_branch_coverage=1 00:16:59.547 --rc genhtml_function_coverage=1 00:16:59.547 --rc genhtml_legend=1 00:16:59.547 --rc geninfo_all_blocks=1 00:16:59.547 --rc geninfo_unexecuted_blocks=1 00:16:59.547 00:16:59.547 ' 00:16:59.547 19:07:16 blockdev_raid5f -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:16:59.547 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:59.547 --rc genhtml_branch_coverage=1 00:16:59.547 --rc genhtml_function_coverage=1 00:16:59.547 --rc genhtml_legend=1 00:16:59.547 --rc geninfo_all_blocks=1 00:16:59.547 --rc geninfo_unexecuted_blocks=1 00:16:59.547 00:16:59.547 ' 00:16:59.547 19:07:16 blockdev_raid5f -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:16:59.547 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:59.547 --rc genhtml_branch_coverage=1 00:16:59.547 --rc genhtml_function_coverage=1 00:16:59.547 --rc genhtml_legend=1 00:16:59.547 --rc geninfo_all_blocks=1 00:16:59.547 --rc geninfo_unexecuted_blocks=1 00:16:59.547 00:16:59.547 ' 00:16:59.547 19:07:16 blockdev_raid5f -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:16:59.547 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:59.547 --rc genhtml_branch_coverage=1 00:16:59.547 --rc genhtml_function_coverage=1 00:16:59.547 --rc genhtml_legend=1 00:16:59.547 --rc geninfo_all_blocks=1 00:16:59.547 --rc geninfo_unexecuted_blocks=1 00:16:59.547 00:16:59.547 ' 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/nbd_common.sh@6 -- # set -e 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@13 -- # conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@20 -- # : 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@707 -- # QOS_DEV_1=Malloc_0 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@708 -- # QOS_DEV_2=Null_1 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@709 -- # QOS_RUN_TIME=5 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@711 -- # uname -s 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@711 -- # '[' Linux = Linux ']' 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@713 -- # PRE_RESERVED_MEM=0 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@719 -- # test_type=raid5f 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@720 -- # crypto_device= 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@721 -- # dek= 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@722 -- # env_ctx= 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@723 -- # wait_for_rpc= 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@724 -- # '[' -n '' ']' 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@727 -- # [[ raid5f == bdev ]] 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@727 -- # [[ raid5f == crypto_* ]] 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@730 -- # start_spdk_tgt 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=96921 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt '' '' 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:16:59.547 19:07:16 blockdev_raid5f -- bdev/blockdev.sh@49 -- # waitforlisten 96921 00:16:59.547 19:07:16 blockdev_raid5f -- common/autotest_common.sh@835 -- # '[' -z 96921 ']' 00:16:59.547 19:07:16 blockdev_raid5f -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:59.547 19:07:16 blockdev_raid5f -- common/autotest_common.sh@840 -- # local max_retries=100 00:16:59.547 19:07:16 blockdev_raid5f -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:59.547 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:59.547 19:07:16 blockdev_raid5f -- common/autotest_common.sh@844 -- # xtrace_disable 00:16:59.547 19:07:16 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:59.547 [2024-12-05 19:07:17.006105] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:16:59.547 [2024-12-05 19:07:17.006315] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96921 ] 00:16:59.807 [2024-12-05 19:07:17.161103] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:59.807 [2024-12-05 19:07:17.187017] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:00.382 19:07:17 blockdev_raid5f -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:00.382 19:07:17 blockdev_raid5f -- common/autotest_common.sh@868 -- # return 0 00:17:00.382 19:07:17 blockdev_raid5f -- bdev/blockdev.sh@731 -- # case "$test_type" in 00:17:00.382 19:07:17 blockdev_raid5f -- bdev/blockdev.sh@763 -- # setup_raid5f_conf 00:17:00.382 19:07:17 blockdev_raid5f -- bdev/blockdev.sh@279 -- # rpc_cmd 00:17:00.382 19:07:17 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.382 19:07:17 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:00.382 Malloc0 00:17:00.382 Malloc1 00:17:00.382 Malloc2 00:17:00.382 19:07:17 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.382 19:07:17 blockdev_raid5f -- bdev/blockdev.sh@774 -- # rpc_cmd bdev_wait_for_examine 00:17:00.382 19:07:17 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.382 19:07:17 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:00.382 19:07:17 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.382 19:07:17 blockdev_raid5f -- bdev/blockdev.sh@777 -- # cat 00:17:00.382 19:07:17 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n accel 00:17:00.382 19:07:17 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.382 19:07:17 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:00.382 19:07:17 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.382 19:07:17 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n bdev 00:17:00.382 19:07:17 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.382 19:07:17 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:00.382 19:07:17 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.642 19:07:17 blockdev_raid5f -- bdev/blockdev.sh@777 -- # rpc_cmd save_subsystem_config -n iobuf 00:17:00.642 19:07:17 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.642 19:07:17 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:00.642 19:07:17 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.642 19:07:17 blockdev_raid5f -- bdev/blockdev.sh@785 -- # mapfile -t bdevs 00:17:00.642 19:07:17 blockdev_raid5f -- bdev/blockdev.sh@785 -- # rpc_cmd bdev_get_bdevs 00:17:00.642 19:07:17 blockdev_raid5f -- bdev/blockdev.sh@785 -- # jq -r '.[] | select(.claimed == false)' 00:17:00.642 19:07:17 blockdev_raid5f -- common/autotest_common.sh@563 -- # xtrace_disable 00:17:00.642 19:07:17 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:00.642 19:07:18 blockdev_raid5f -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:17:00.642 19:07:18 blockdev_raid5f -- bdev/blockdev.sh@786 -- # mapfile -t bdevs_name 00:17:00.642 19:07:18 blockdev_raid5f -- bdev/blockdev.sh@786 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "5a88721a-cabe-44a2-84bb-a955fced1696"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "5a88721a-cabe-44a2-84bb-a955fced1696",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "5a88721a-cabe-44a2-84bb-a955fced1696",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "30e578da-8384-4677-8b9f-1f95a137fcfc",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "0d150e0e-505f-4713-95db-df3972b00ff6",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "8426dffe-80a4-43ae-8706-c9a0af3ba408",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:17:00.642 19:07:18 blockdev_raid5f -- bdev/blockdev.sh@786 -- # jq -r .name 00:17:00.642 19:07:18 blockdev_raid5f -- bdev/blockdev.sh@787 -- # bdev_list=("${bdevs_name[@]}") 00:17:00.642 19:07:18 blockdev_raid5f -- bdev/blockdev.sh@789 -- # hello_world_bdev=raid5f 00:17:00.642 19:07:18 blockdev_raid5f -- bdev/blockdev.sh@790 -- # trap - SIGINT SIGTERM EXIT 00:17:00.642 19:07:18 blockdev_raid5f -- bdev/blockdev.sh@791 -- # killprocess 96921 00:17:00.642 19:07:18 blockdev_raid5f -- common/autotest_common.sh@954 -- # '[' -z 96921 ']' 00:17:00.642 19:07:18 blockdev_raid5f -- common/autotest_common.sh@958 -- # kill -0 96921 00:17:00.642 19:07:18 blockdev_raid5f -- common/autotest_common.sh@959 -- # uname 00:17:00.642 19:07:18 blockdev_raid5f -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:00.642 19:07:18 blockdev_raid5f -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 96921 00:17:00.642 killing process with pid 96921 00:17:00.642 19:07:18 blockdev_raid5f -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:00.642 19:07:18 blockdev_raid5f -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:00.642 19:07:18 blockdev_raid5f -- common/autotest_common.sh@972 -- # echo 'killing process with pid 96921' 00:17:00.642 19:07:18 blockdev_raid5f -- common/autotest_common.sh@973 -- # kill 96921 00:17:00.642 19:07:18 blockdev_raid5f -- common/autotest_common.sh@978 -- # wait 96921 00:17:01.211 19:07:18 blockdev_raid5f -- bdev/blockdev.sh@795 -- # trap cleanup SIGINT SIGTERM EXIT 00:17:01.211 19:07:18 blockdev_raid5f -- bdev/blockdev.sh@797 -- # run_test bdev_hello_world /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:17:01.211 19:07:18 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 7 -le 1 ']' 00:17:01.211 19:07:18 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:01.211 19:07:18 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:01.211 ************************************ 00:17:01.211 START TEST bdev_hello_world 00:17:01.211 ************************************ 00:17:01.211 19:07:18 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:17:01.211 [2024-12-05 19:07:18.584020] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:17:01.211 [2024-12-05 19:07:18.584232] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96954 ] 00:17:01.211 [2024-12-05 19:07:18.737360] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:01.211 [2024-12-05 19:07:18.763929] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:01.471 [2024-12-05 19:07:18.942550] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:17:01.471 [2024-12-05 19:07:18.942695] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev raid5f 00:17:01.471 [2024-12-05 19:07:18.942727] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:17:01.471 [2024-12-05 19:07:18.943056] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:17:01.471 [2024-12-05 19:07:18.943236] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:17:01.471 [2024-12-05 19:07:18.943294] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:17:01.471 [2024-12-05 19:07:18.943368] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:17:01.471 00:17:01.471 [2024-12-05 19:07:18.943420] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:17:01.732 00:17:01.732 ************************************ 00:17:01.732 END TEST bdev_hello_world 00:17:01.732 ************************************ 00:17:01.732 real 0m0.656s 00:17:01.732 user 0m0.361s 00:17:01.732 sys 0m0.189s 00:17:01.732 19:07:19 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:01.732 19:07:19 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:17:01.732 19:07:19 blockdev_raid5f -- bdev/blockdev.sh@798 -- # run_test bdev_bounds bdev_bounds '' 00:17:01.732 19:07:19 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:17:01.732 19:07:19 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:01.732 19:07:19 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:01.732 ************************************ 00:17:01.732 START TEST bdev_bounds 00:17:01.732 ************************************ 00:17:01.732 19:07:19 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1129 -- # bdev_bounds '' 00:17:01.732 19:07:19 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@289 -- # bdevio_pid=96974 00:17:01.732 19:07:19 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@288 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:17:01.732 19:07:19 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@290 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:17:01.732 19:07:19 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@291 -- # echo 'Process bdevio pid: 96974' 00:17:01.732 Process bdevio pid: 96974 00:17:01.732 19:07:19 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@292 -- # waitforlisten 96974 00:17:01.732 19:07:19 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@835 -- # '[' -z 96974 ']' 00:17:01.732 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:01.732 19:07:19 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:01.732 19:07:19 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:01.732 19:07:19 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:01.732 19:07:19 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:01.732 19:07:19 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:17:01.994 [2024-12-05 19:07:19.329338] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:17:01.994 [2024-12-05 19:07:19.329556] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96974 ] 00:17:01.994 [2024-12-05 19:07:19.485719] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:17:01.994 [2024-12-05 19:07:19.514040] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:17:01.994 [2024-12-05 19:07:19.514197] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:01.994 [2024-12-05 19:07:19.514245] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:17:02.933 19:07:20 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:02.933 19:07:20 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@868 -- # return 0 00:17:02.933 19:07:20 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@293 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/tests.py perform_tests 00:17:02.933 I/O targets: 00:17:02.933 raid5f: 131072 blocks of 512 bytes (64 MiB) 00:17:02.933 00:17:02.933 00:17:02.933 CUnit - A unit testing framework for C - Version 2.1-3 00:17:02.933 http://cunit.sourceforge.net/ 00:17:02.933 00:17:02.933 00:17:02.933 Suite: bdevio tests on: raid5f 00:17:02.933 Test: blockdev write read block ...passed 00:17:02.933 Test: blockdev write zeroes read block ...passed 00:17:02.933 Test: blockdev write zeroes read no split ...passed 00:17:02.933 Test: blockdev write zeroes read split ...passed 00:17:02.933 Test: blockdev write zeroes read split partial ...passed 00:17:02.933 Test: blockdev reset ...passed 00:17:02.933 Test: blockdev write read 8 blocks ...passed 00:17:02.933 Test: blockdev write read size > 128k ...passed 00:17:02.933 Test: blockdev write read invalid size ...passed 00:17:02.933 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:17:02.933 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:17:02.933 Test: blockdev write read max offset ...passed 00:17:02.933 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:17:02.933 Test: blockdev writev readv 8 blocks ...passed 00:17:02.933 Test: blockdev writev readv 30 x 1block ...passed 00:17:02.933 Test: blockdev writev readv block ...passed 00:17:02.933 Test: blockdev writev readv size > 128k ...passed 00:17:02.933 Test: blockdev writev readv size > 128k in two iovs ...passed 00:17:02.933 Test: blockdev comparev and writev ...passed 00:17:02.933 Test: blockdev nvme passthru rw ...passed 00:17:02.933 Test: blockdev nvme passthru vendor specific ...passed 00:17:02.933 Test: blockdev nvme admin passthru ...passed 00:17:02.933 Test: blockdev copy ...passed 00:17:02.933 00:17:02.933 Run Summary: Type Total Ran Passed Failed Inactive 00:17:02.933 suites 1 1 n/a 0 0 00:17:02.933 tests 23 23 23 0 0 00:17:02.933 asserts 130 130 130 0 n/a 00:17:02.933 00:17:02.933 Elapsed time = 0.329 seconds 00:17:02.933 0 00:17:02.933 19:07:20 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@294 -- # killprocess 96974 00:17:02.933 19:07:20 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@954 -- # '[' -z 96974 ']' 00:17:02.933 19:07:20 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@958 -- # kill -0 96974 00:17:02.933 19:07:20 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@959 -- # uname 00:17:02.933 19:07:20 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:02.933 19:07:20 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 96974 00:17:02.933 19:07:20 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:02.933 19:07:20 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:02.933 19:07:20 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@972 -- # echo 'killing process with pid 96974' 00:17:02.933 killing process with pid 96974 00:17:02.933 19:07:20 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@973 -- # kill 96974 00:17:02.933 19:07:20 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@978 -- # wait 96974 00:17:03.193 19:07:20 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@295 -- # trap - SIGINT SIGTERM EXIT 00:17:03.193 00:17:03.193 real 0m1.463s 00:17:03.193 user 0m3.517s 00:17:03.193 sys 0m0.371s 00:17:03.193 19:07:20 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:03.193 19:07:20 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:17:03.193 ************************************ 00:17:03.193 END TEST bdev_bounds 00:17:03.193 ************************************ 00:17:03.454 19:07:20 blockdev_raid5f -- bdev/blockdev.sh@799 -- # run_test bdev_nbd nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:17:03.454 19:07:20 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:17:03.454 19:07:20 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:03.454 19:07:20 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:03.454 ************************************ 00:17:03.454 START TEST bdev_nbd 00:17:03.454 ************************************ 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1129 -- # nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # uname -s 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # [[ Linux == Linux ]] 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@301 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@302 -- # local conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # bdev_all=('raid5f') 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # local bdev_all 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@304 -- # local bdev_num=1 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@308 -- # [[ -e /sys/module/nbd ]] 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # local nbd_all 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@311 -- # bdev_num=1 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # nbd_list=('/dev/nbd0') 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # local nbd_list 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # bdev_list=('raid5f') 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # local bdev_list 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@317 -- # nbd_pid=97022 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@318 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@316 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@319 -- # waitforlisten 97022 /var/tmp/spdk-nbd.sock 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@835 -- # '[' -z 97022 ']' 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:17:03.454 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@840 -- # local max_retries=100 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@844 -- # xtrace_disable 00:17:03.454 19:07:20 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:17:03.454 [2024-12-05 19:07:20.879398] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:17:03.454 [2024-12-05 19:07:20.879601] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:03.715 [2024-12-05 19:07:21.038043] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:03.715 [2024-12-05 19:07:21.063253] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:04.285 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:17:04.285 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@868 -- # return 0 00:17:04.285 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@321 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock raid5f 00:17:04.285 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:04.285 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('raid5f') 00:17:04.285 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:17:04.285 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock raid5f 00:17:04.285 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:04.285 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('raid5f') 00:17:04.285 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:17:04.285 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:17:04.285 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:17:04.285 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:17:04.285 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:17:04.285 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # local i 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@877 -- # break 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:04.545 1+0 records in 00:17:04.545 1+0 records out 00:17:04.545 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00045445 s, 9.0 MB/s 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # size=4096 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@893 -- # return 0 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:17:04.545 19:07:21 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:04.805 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:17:04.805 { 00:17:04.805 "nbd_device": "/dev/nbd0", 00:17:04.805 "bdev_name": "raid5f" 00:17:04.805 } 00:17:04.805 ]' 00:17:04.805 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:17:04.805 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:17:04.805 { 00:17:04.805 "nbd_device": "/dev/nbd0", 00:17:04.805 "bdev_name": "raid5f" 00:17:04.805 } 00:17:04.805 ]' 00:17:04.805 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:17:04.805 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:04.805 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:04.805 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:17:04.805 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:04.805 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:17:04.805 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:04.805 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:17:05.065 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:05.065 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:05.065 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:05.065 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:05.065 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:05.065 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:05.065 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:17:05.065 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:17:05.065 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:05.065 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:05.065 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@322 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('raid5f') 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0') 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('raid5f') 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:17:05.325 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f /dev/nbd0 00:17:05.585 /dev/nbd0 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # local i 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@877 -- # break 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:05.585 1+0 records in 00:17:05.585 1+0 records out 00:17:05.585 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000347555 s, 11.8 MB/s 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@890 -- # size=4096 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@891 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@893 -- # return 0 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:05.585 19:07:22 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:05.845 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:17:05.845 { 00:17:05.845 "nbd_device": "/dev/nbd0", 00:17:05.845 "bdev_name": "raid5f" 00:17:05.845 } 00:17:05.845 ]' 00:17:05.845 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:05.845 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:17:05.845 { 00:17:05.845 "nbd_device": "/dev/nbd0", 00:17:05.845 "bdev_name": "raid5f" 00:17:05.845 } 00:17:05.845 ]' 00:17:05.845 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=1 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 1 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=1 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 1 -ne 1 ']' 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify /dev/nbd0 write 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:17:05.846 256+0 records in 00:17:05.846 256+0 records out 00:17:05.846 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0128803 s, 81.4 MB/s 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:17:05.846 256+0 records in 00:17:05.846 256+0 records out 00:17:05.846 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0254668 s, 41.2 MB/s 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify /dev/nbd0 verify 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest /dev/nbd0 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:05.846 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:17:06.105 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:06.105 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:06.105 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:06.106 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:06.106 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:06.106 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:06.106 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:17:06.106 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:17:06.106 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:06.106 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:06.106 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd=/dev/nbd0 00:17:06.365 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@134 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:17:06.625 malloc_lvol_verify 00:17:06.625 19:07:23 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@135 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:17:06.625 6954ab57-0c16-454e-8b9b-84cd2ed779fe 00:17:06.625 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@136 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:17:06.884 caaec608-d3f1-4f7e-819f-538aa9d65165 00:17:06.884 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@137 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:17:07.145 /dev/nbd0 00:17:07.145 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@139 -- # wait_for_nbd_set_capacity /dev/nbd0 00:17:07.145 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@146 -- # local nbd=nbd0 00:17:07.145 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@148 -- # [[ -e /sys/block/nbd0/size ]] 00:17:07.145 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@150 -- # (( 8192 == 0 )) 00:17:07.145 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs.ext4 /dev/nbd0 00:17:07.145 Discarding device blocks: 0/4096mke2fs 1.47.0 (5-Feb-2023) 00:17:07.145  done 00:17:07.145 Creating filesystem with 4096 1k blocks and 1024 inodes 00:17:07.145 00:17:07.145 Allocating group tables: 0/1 done 00:17:07.145 Writing inode tables: 0/1 done 00:17:07.145 Creating journal (1024 blocks): done 00:17:07.145 Writing superblocks and filesystem accounting information: 0/1 done 00:17:07.145 00:17:07.145 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:07.145 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:07.145 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:17:07.145 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:07.145 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:17:07.145 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:07.145 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@325 -- # killprocess 97022 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@954 -- # '[' -z 97022 ']' 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@958 -- # kill -0 97022 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@959 -- # uname 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 97022 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:17:07.405 killing process with pid 97022 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@972 -- # echo 'killing process with pid 97022' 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@973 -- # kill 97022 00:17:07.405 19:07:24 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@978 -- # wait 97022 00:17:07.666 19:07:25 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@326 -- # trap - SIGINT SIGTERM EXIT 00:17:07.666 00:17:07.666 real 0m4.365s 00:17:07.666 user 0m6.317s 00:17:07.666 sys 0m1.312s 00:17:07.666 19:07:25 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:07.666 19:07:25 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:17:07.666 ************************************ 00:17:07.666 END TEST bdev_nbd 00:17:07.666 ************************************ 00:17:07.666 19:07:25 blockdev_raid5f -- bdev/blockdev.sh@800 -- # [[ y == y ]] 00:17:07.666 19:07:25 blockdev_raid5f -- bdev/blockdev.sh@801 -- # '[' raid5f = nvme ']' 00:17:07.666 19:07:25 blockdev_raid5f -- bdev/blockdev.sh@801 -- # '[' raid5f = gpt ']' 00:17:07.666 19:07:25 blockdev_raid5f -- bdev/blockdev.sh@805 -- # run_test bdev_fio fio_test_suite '' 00:17:07.666 19:07:25 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:17:07.666 19:07:25 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:07.666 19:07:25 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:07.666 ************************************ 00:17:07.666 START TEST bdev_fio 00:17:07.666 ************************************ 00:17:07.666 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1129 -- # fio_test_suite '' 00:17:07.666 19:07:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@330 -- # local env_context 00:17:07.666 19:07:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@334 -- # pushd /home/vagrant/spdk_repo/spdk/test/bdev 00:17:07.666 /home/vagrant/spdk_repo/spdk/test/bdev /home/vagrant/spdk_repo/spdk 00:17:07.666 19:07:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@335 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:17:07.666 19:07:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # echo '' 00:17:07.666 19:07:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # sed s/--env-context=// 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # env_context= 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@339 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio verify AIO '' 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local workload=verify 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local bdev_type=AIO 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1287 -- # local env_context= 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # local fio_dir=/usr/src/fio 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1290 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -z verify ']' 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # '[' -n '' ']' 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1305 -- # cat 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1317 -- # '[' verify == verify ']' 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1318 -- # cat 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1327 -- # '[' AIO == AIO ']' 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # /usr/src/fio/fio --version 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1329 -- # echo serialize_overlap=1 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@340 -- # for b in "${bdevs_name[@]}" 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@341 -- # echo '[job_raid5f]' 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@342 -- # echo filename=raid5f 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@346 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json' 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@348 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1105 -- # '[' 11 -le 1 ']' 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:17:07.926 ************************************ 00:17:07.926 START TEST bdev_fio_rw_verify 00:17:07.926 ************************************ 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1129 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1360 -- # fio_plugin /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # local sanitizers 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # local plugin=/home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # shift 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1347 -- # local asan_lib= 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # ldd /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # grep libasan 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1349 -- # asan_lib=/usr/lib64/libasan.so.8 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1350 -- # [[ -n /usr/lib64/libasan.so.8 ]] 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1351 -- # break 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # LD_PRELOAD='/usr/lib64/libasan.so.8 /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev' 00:17:07.926 19:07:25 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:08.186 job_raid5f: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:17:08.186 fio-3.35 00:17:08.186 Starting 1 thread 00:17:20.406 00:17:20.406 job_raid5f: (groupid=0, jobs=1): err= 0: pid=97176: Thu Dec 5 19:07:36 2024 00:17:20.406 read: IOPS=12.5k, BW=48.8MiB/s (51.2MB/s)(488MiB/10001msec) 00:17:20.406 slat (usec): min=17, max=240, avg=19.33, stdev= 1.87 00:17:20.406 clat (usec): min=10, max=490, avg=129.33, stdev=45.45 00:17:20.406 lat (usec): min=30, max=510, avg=148.66, stdev=45.63 00:17:20.406 clat percentiles (usec): 00:17:20.406 | 50.000th=[ 135], 99.000th=[ 206], 99.900th=[ 231], 99.990th=[ 334], 00:17:20.406 | 99.999th=[ 486] 00:17:20.406 write: IOPS=13.1k, BW=51.2MiB/s (53.7MB/s)(505MiB/9875msec); 0 zone resets 00:17:20.406 slat (usec): min=7, max=270, avg=15.93, stdev= 4.08 00:17:20.406 clat (usec): min=57, max=2178, avg=294.32, stdev=51.75 00:17:20.406 lat (usec): min=71, max=2448, avg=310.25, stdev=53.60 00:17:20.406 clat percentiles (usec): 00:17:20.406 | 50.000th=[ 297], 99.000th=[ 371], 99.900th=[ 930], 99.990th=[ 1565], 00:17:20.406 | 99.999th=[ 2089] 00:17:20.406 bw ( KiB/s): min=49328, max=54744, per=98.85%, avg=51794.11, stdev=1491.04, samples=19 00:17:20.406 iops : min=12332, max=13686, avg=12948.53, stdev=372.76, samples=19 00:17:20.406 lat (usec) : 20=0.01%, 50=0.01%, 100=17.04%, 250=39.16%, 500=43.61% 00:17:20.406 lat (usec) : 750=0.10%, 1000=0.05% 00:17:20.406 lat (msec) : 2=0.04%, 4=0.01% 00:17:20.406 cpu : usr=98.84%, sys=0.45%, ctx=30, majf=0, minf=13302 00:17:20.406 IO depths : 1=7.6%, 2=19.8%, 4=55.2%, 8=17.4%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:20.406 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:20.406 complete : 0=0.0%, 4=90.0%, 8=10.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:20.406 issued rwts: total=124908,129354,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:20.406 latency : target=0, window=0, percentile=100.00%, depth=8 00:17:20.406 00:17:20.406 Run status group 0 (all jobs): 00:17:20.406 READ: bw=48.8MiB/s (51.2MB/s), 48.8MiB/s-48.8MiB/s (51.2MB/s-51.2MB/s), io=488MiB (512MB), run=10001-10001msec 00:17:20.406 WRITE: bw=51.2MiB/s (53.7MB/s), 51.2MiB/s-51.2MiB/s (53.7MB/s-53.7MB/s), io=505MiB (530MB), run=9875-9875msec 00:17:20.406 ----------------------------------------------------- 00:17:20.406 Suppressions used: 00:17:20.406 count bytes template 00:17:20.406 1 7 /usr/src/fio/parse.c 00:17:20.406 561 53856 /usr/src/fio/iolog.c 00:17:20.406 1 8 libtcmalloc_minimal.so 00:17:20.406 1 904 libcrypto.so 00:17:20.406 ----------------------------------------------------- 00:17:20.406 00:17:20.406 00:17:20.406 real 0m11.203s 00:17:20.406 user 0m11.482s 00:17:20.406 sys 0m0.462s 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:17:20.406 ************************************ 00:17:20.406 END TEST bdev_fio_rw_verify 00:17:20.406 ************************************ 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@349 -- # rm -f 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@350 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@353 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio trim '' '' 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1285 -- # local workload=trim 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # local bdev_type= 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1287 -- # local env_context= 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1288 -- # local fio_dir=/usr/src/fio 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1290 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -z trim ']' 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # '[' -n '' ']' 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1303 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1305 -- # cat 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1317 -- # '[' trim == verify ']' 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1332 -- # '[' trim == trim ']' 00:17:20.406 19:07:36 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1333 -- # echo rw=trimwrite 00:17:20.407 19:07:36 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "5a88721a-cabe-44a2-84bb-a955fced1696"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "5a88721a-cabe-44a2-84bb-a955fced1696",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "5a88721a-cabe-44a2-84bb-a955fced1696",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "30e578da-8384-4677-8b9f-1f95a137fcfc",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "0d150e0e-505f-4713-95db-df3972b00ff6",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "8426dffe-80a4-43ae-8706-c9a0af3ba408",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:17:20.407 19:07:36 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:17:20.407 19:07:36 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # [[ -n '' ]] 00:17:20.407 19:07:36 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@360 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:20.407 19:07:36 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@361 -- # popd 00:17:20.407 /home/vagrant/spdk_repo/spdk 00:17:20.407 19:07:36 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@362 -- # trap - SIGINT SIGTERM EXIT 00:17:20.407 19:07:36 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@363 -- # return 0 00:17:20.407 00:17:20.407 real 0m11.506s 00:17:20.407 user 0m11.615s 00:17:20.407 sys 0m0.598s 00:17:20.407 19:07:36 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:20.407 19:07:36 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:17:20.407 ************************************ 00:17:20.407 END TEST bdev_fio 00:17:20.407 ************************************ 00:17:20.407 19:07:36 blockdev_raid5f -- bdev/blockdev.sh@812 -- # trap cleanup SIGINT SIGTERM EXIT 00:17:20.407 19:07:36 blockdev_raid5f -- bdev/blockdev.sh@814 -- # run_test bdev_verify /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:17:20.407 19:07:36 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 16 -le 1 ']' 00:17:20.407 19:07:36 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:20.407 19:07:36 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:20.407 ************************************ 00:17:20.407 START TEST bdev_verify 00:17:20.407 ************************************ 00:17:20.407 19:07:36 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:17:20.407 [2024-12-05 19:07:36.893716] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:17:20.407 [2024-12-05 19:07:36.893855] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97270 ] 00:17:20.407 [2024-12-05 19:07:37.050028] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:20.407 [2024-12-05 19:07:37.081026] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:20.407 [2024-12-05 19:07:37.081198] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:17:20.407 Running I/O for 5 seconds... 00:17:21.915 11853.00 IOPS, 46.30 MiB/s [2024-12-05T19:07:40.412Z] 11294.00 IOPS, 44.12 MiB/s [2024-12-05T19:07:41.349Z] 11107.67 IOPS, 43.39 MiB/s [2024-12-05T19:07:42.728Z] 10988.25 IOPS, 42.92 MiB/s [2024-12-05T19:07:42.728Z] 10961.60 IOPS, 42.82 MiB/s 00:17:25.169 Latency(us) 00:17:25.169 [2024-12-05T19:07:42.728Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:25.169 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:17:25.169 Verification LBA range: start 0x0 length 0x2000 00:17:25.169 raid5f : 5.02 4544.42 17.75 0.00 0.00 42227.67 139.51 30907.81 00:17:25.169 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:17:25.169 Verification LBA range: start 0x2000 length 0x2000 00:17:25.169 raid5f : 5.01 6402.09 25.01 0.00 0.00 30003.08 178.86 36402.53 00:17:25.169 [2024-12-05T19:07:42.728Z] =================================================================================================================== 00:17:25.169 [2024-12-05T19:07:42.728Z] Total : 10946.51 42.76 0.00 0.00 35083.53 139.51 36402.53 00:17:25.169 00:17:25.169 real 0m5.726s 00:17:25.169 user 0m10.664s 00:17:25.169 sys 0m0.247s 00:17:25.169 19:07:42 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:25.169 19:07:42 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:17:25.169 ************************************ 00:17:25.169 END TEST bdev_verify 00:17:25.169 ************************************ 00:17:25.169 19:07:42 blockdev_raid5f -- bdev/blockdev.sh@815 -- # run_test bdev_verify_big_io /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:17:25.169 19:07:42 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 16 -le 1 ']' 00:17:25.169 19:07:42 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:25.169 19:07:42 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:25.169 ************************************ 00:17:25.169 START TEST bdev_verify_big_io 00:17:25.169 ************************************ 00:17:25.169 19:07:42 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:17:25.169 [2024-12-05 19:07:42.684198] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:17:25.169 [2024-12-05 19:07:42.684312] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97316 ] 00:17:25.428 [2024-12-05 19:07:42.840764] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:25.428 [2024-12-05 19:07:42.874762] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:25.428 [2024-12-05 19:07:42.874858] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:17:25.687 Running I/O for 5 seconds... 00:17:27.567 633.00 IOPS, 39.56 MiB/s [2024-12-05T19:07:46.508Z] 761.00 IOPS, 47.56 MiB/s [2024-12-05T19:07:47.450Z] 781.33 IOPS, 48.83 MiB/s [2024-12-05T19:07:48.391Z] 777.00 IOPS, 48.56 MiB/s [2024-12-05T19:07:48.391Z] 787.00 IOPS, 49.19 MiB/s 00:17:30.832 Latency(us) 00:17:30.832 [2024-12-05T19:07:48.391Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:30.832 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:17:30.832 Verification LBA range: start 0x0 length 0x200 00:17:30.832 raid5f : 5.20 341.79 21.36 0.00 0.00 9246354.93 215.53 384630.50 00:17:30.832 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:17:30.832 Verification LBA range: start 0x200 length 0x200 00:17:30.832 raid5f : 5.20 451.60 28.22 0.00 0.00 7037163.32 261.14 309535.97 00:17:30.832 [2024-12-05T19:07:48.391Z] =================================================================================================================== 00:17:30.832 [2024-12-05T19:07:48.391Z] Total : 793.39 49.59 0.00 0.00 7989161.06 215.53 384630.50 00:17:31.091 00:17:31.091 real 0m5.902s 00:17:31.091 user 0m11.033s 00:17:31.091 sys 0m0.236s 00:17:31.091 19:07:48 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:31.091 19:07:48 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:17:31.091 ************************************ 00:17:31.091 END TEST bdev_verify_big_io 00:17:31.091 ************************************ 00:17:31.092 19:07:48 blockdev_raid5f -- bdev/blockdev.sh@816 -- # run_test bdev_write_zeroes /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:31.092 19:07:48 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:17:31.092 19:07:48 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:31.092 19:07:48 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:31.092 ************************************ 00:17:31.092 START TEST bdev_write_zeroes 00:17:31.092 ************************************ 00:17:31.092 19:07:48 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:31.358 [2024-12-05 19:07:48.664346] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:17:31.358 [2024-12-05 19:07:48.665057] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97363 ] 00:17:31.358 [2024-12-05 19:07:48.820044] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:31.358 [2024-12-05 19:07:48.849428] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:31.635 Running I/O for 1 seconds... 00:17:32.603 29727.00 IOPS, 116.12 MiB/s 00:17:32.603 Latency(us) 00:17:32.603 [2024-12-05T19:07:50.162Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:32.603 Job: raid5f (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:17:32.603 raid5f : 1.01 29687.03 115.96 0.00 0.00 4299.05 1409.45 5924.00 00:17:32.603 [2024-12-05T19:07:50.162Z] =================================================================================================================== 00:17:32.603 [2024-12-05T19:07:50.162Z] Total : 29687.03 115.96 0.00 0.00 4299.05 1409.45 5924.00 00:17:32.864 00:17:32.864 real 0m1.687s 00:17:32.864 user 0m1.369s 00:17:32.864 sys 0m0.206s 00:17:32.864 19:07:50 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:32.864 19:07:50 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:17:32.864 ************************************ 00:17:32.864 END TEST bdev_write_zeroes 00:17:32.864 ************************************ 00:17:32.864 19:07:50 blockdev_raid5f -- bdev/blockdev.sh@819 -- # run_test bdev_json_nonenclosed /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:32.864 19:07:50 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:17:32.864 19:07:50 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:32.864 19:07:50 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:32.864 ************************************ 00:17:32.864 START TEST bdev_json_nonenclosed 00:17:32.864 ************************************ 00:17:32.864 19:07:50 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:32.864 [2024-12-05 19:07:50.419286] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:17:32.864 [2024-12-05 19:07:50.419409] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97392 ] 00:17:33.128 [2024-12-05 19:07:50.576350] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:33.128 [2024-12-05 19:07:50.607307] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:33.128 [2024-12-05 19:07:50.607421] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:17:33.128 [2024-12-05 19:07:50.607450] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:17:33.128 [2024-12-05 19:07:50.607461] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:17:33.388 00:17:33.388 real 0m0.359s 00:17:33.388 user 0m0.139s 00:17:33.388 sys 0m0.117s 00:17:33.388 19:07:50 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:33.388 19:07:50 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:17:33.388 ************************************ 00:17:33.388 END TEST bdev_json_nonenclosed 00:17:33.388 ************************************ 00:17:33.388 19:07:50 blockdev_raid5f -- bdev/blockdev.sh@822 -- # run_test bdev_json_nonarray /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:33.388 19:07:50 blockdev_raid5f -- common/autotest_common.sh@1105 -- # '[' 13 -le 1 ']' 00:17:33.388 19:07:50 blockdev_raid5f -- common/autotest_common.sh@1111 -- # xtrace_disable 00:17:33.388 19:07:50 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:33.388 ************************************ 00:17:33.388 START TEST bdev_json_nonarray 00:17:33.388 ************************************ 00:17:33.388 19:07:50 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1129 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:33.388 [2024-12-05 19:07:50.856486] Starting SPDK v25.01-pre git sha1 8d3947977 / DPDK 22.11.4 initialization... 00:17:33.388 [2024-12-05 19:07:50.856626] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97412 ] 00:17:33.647 [2024-12-05 19:07:51.012720] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:33.648 [2024-12-05 19:07:51.038432] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:33.648 [2024-12-05 19:07:51.038542] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:17:33.648 [2024-12-05 19:07:51.038562] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:17:33.648 [2024-12-05 19:07:51.038574] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:17:33.648 00:17:33.648 real 0m0.355s 00:17:33.648 user 0m0.137s 00:17:33.648 sys 0m0.114s 00:17:33.648 19:07:51 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:33.648 19:07:51 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:17:33.648 ************************************ 00:17:33.648 END TEST bdev_json_nonarray 00:17:33.648 ************************************ 00:17:33.648 19:07:51 blockdev_raid5f -- bdev/blockdev.sh@824 -- # [[ raid5f == bdev ]] 00:17:33.648 19:07:51 blockdev_raid5f -- bdev/blockdev.sh@832 -- # [[ raid5f == gpt ]] 00:17:33.648 19:07:51 blockdev_raid5f -- bdev/blockdev.sh@836 -- # [[ raid5f == crypto_sw ]] 00:17:33.648 19:07:51 blockdev_raid5f -- bdev/blockdev.sh@848 -- # trap - SIGINT SIGTERM EXIT 00:17:33.648 19:07:51 blockdev_raid5f -- bdev/blockdev.sh@849 -- # cleanup 00:17:33.648 19:07:51 blockdev_raid5f -- bdev/blockdev.sh@23 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/aiofile 00:17:33.648 19:07:51 blockdev_raid5f -- bdev/blockdev.sh@24 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:17:33.648 19:07:51 blockdev_raid5f -- bdev/blockdev.sh@26 -- # [[ raid5f == rbd ]] 00:17:33.648 19:07:51 blockdev_raid5f -- bdev/blockdev.sh@30 -- # [[ raid5f == daos ]] 00:17:33.648 19:07:51 blockdev_raid5f -- bdev/blockdev.sh@34 -- # [[ raid5f = \g\p\t ]] 00:17:33.648 19:07:51 blockdev_raid5f -- bdev/blockdev.sh@40 -- # [[ raid5f == xnvme ]] 00:17:33.648 00:17:33.648 real 0m34.538s 00:17:33.648 user 0m47.104s 00:17:33.648 sys 0m4.474s 00:17:33.648 19:07:51 blockdev_raid5f -- common/autotest_common.sh@1130 -- # xtrace_disable 00:17:33.648 19:07:51 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:33.648 ************************************ 00:17:33.648 END TEST blockdev_raid5f 00:17:33.648 ************************************ 00:17:33.907 19:07:51 -- spdk/autotest.sh@194 -- # uname -s 00:17:33.907 19:07:51 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:17:33.907 19:07:51 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:17:33.907 19:07:51 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:17:33.907 19:07:51 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@256 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@260 -- # timing_exit lib 00:17:33.907 19:07:51 -- common/autotest_common.sh@732 -- # xtrace_disable 00:17:33.907 19:07:51 -- common/autotest_common.sh@10 -- # set +x 00:17:33.907 19:07:51 -- spdk/autotest.sh@262 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@267 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@276 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@319 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@324 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@333 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@350 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@359 -- # '[' 0 -eq 1 ']' 00:17:33.907 19:07:51 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:17:33.907 19:07:51 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:17:33.907 19:07:51 -- spdk/autotest.sh@374 -- # [[ 0 -eq 1 ]] 00:17:33.907 19:07:51 -- spdk/autotest.sh@378 -- # [[ '' -eq 1 ]] 00:17:33.907 19:07:51 -- spdk/autotest.sh@385 -- # trap - SIGINT SIGTERM EXIT 00:17:33.907 19:07:51 -- spdk/autotest.sh@387 -- # timing_enter post_cleanup 00:17:33.907 19:07:51 -- common/autotest_common.sh@726 -- # xtrace_disable 00:17:33.907 19:07:51 -- common/autotest_common.sh@10 -- # set +x 00:17:33.907 19:07:51 -- spdk/autotest.sh@388 -- # autotest_cleanup 00:17:33.907 19:07:51 -- common/autotest_common.sh@1396 -- # local autotest_es=0 00:17:33.907 19:07:51 -- common/autotest_common.sh@1397 -- # xtrace_disable 00:17:33.907 19:07:51 -- common/autotest_common.sh@10 -- # set +x 00:17:36.447 INFO: APP EXITING 00:17:36.447 INFO: killing all VMs 00:17:36.447 INFO: killing vhost app 00:17:36.447 INFO: EXIT DONE 00:17:36.707 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:17:36.707 Waiting for block devices as requested 00:17:36.707 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:17:36.967 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:17:37.906 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:17:37.906 Cleaning 00:17:37.906 Removing: /var/run/dpdk/spdk0/config 00:17:37.906 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:17:37.906 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:17:37.906 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:17:37.906 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:17:37.906 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:17:37.906 Removing: /var/run/dpdk/spdk0/hugepage_info 00:17:37.906 Removing: /dev/shm/spdk_tgt_trace.pid68836 00:17:37.906 Removing: /var/run/dpdk/spdk0 00:17:37.906 Removing: /var/run/dpdk/spdk_pid68673 00:17:37.907 Removing: /var/run/dpdk/spdk_pid68836 00:17:37.907 Removing: /var/run/dpdk/spdk_pid69038 00:17:37.907 Removing: /var/run/dpdk/spdk_pid69125 00:17:37.907 Removing: /var/run/dpdk/spdk_pid69154 00:17:37.907 Removing: /var/run/dpdk/spdk_pid69262 00:17:37.907 Removing: /var/run/dpdk/spdk_pid69278 00:17:37.907 Removing: /var/run/dpdk/spdk_pid69466 00:17:37.907 Removing: /var/run/dpdk/spdk_pid69545 00:17:37.907 Removing: /var/run/dpdk/spdk_pid69619 00:17:37.907 Removing: /var/run/dpdk/spdk_pid69719 00:17:37.907 Removing: /var/run/dpdk/spdk_pid69805 00:17:37.907 Removing: /var/run/dpdk/spdk_pid69839 00:17:37.907 Removing: /var/run/dpdk/spdk_pid69870 00:17:37.907 Removing: /var/run/dpdk/spdk_pid69946 00:17:37.907 Removing: /var/run/dpdk/spdk_pid70058 00:17:37.907 Removing: /var/run/dpdk/spdk_pid70483 00:17:37.907 Removing: /var/run/dpdk/spdk_pid70530 00:17:37.907 Removing: /var/run/dpdk/spdk_pid70577 00:17:37.907 Removing: /var/run/dpdk/spdk_pid70593 00:17:37.907 Removing: /var/run/dpdk/spdk_pid70658 00:17:37.907 Removing: /var/run/dpdk/spdk_pid70669 00:17:37.907 Removing: /var/run/dpdk/spdk_pid70738 00:17:37.907 Removing: /var/run/dpdk/spdk_pid70754 00:17:37.907 Removing: /var/run/dpdk/spdk_pid70796 00:17:37.907 Removing: /var/run/dpdk/spdk_pid70814 00:17:37.907 Removing: /var/run/dpdk/spdk_pid70856 00:17:37.907 Removing: /var/run/dpdk/spdk_pid70874 00:17:37.907 Removing: /var/run/dpdk/spdk_pid71012 00:17:37.907 Removing: /var/run/dpdk/spdk_pid71045 00:17:37.907 Removing: /var/run/dpdk/spdk_pid71133 00:17:38.166 Removing: /var/run/dpdk/spdk_pid72291 00:17:38.166 Removing: /var/run/dpdk/spdk_pid72486 00:17:38.166 Removing: /var/run/dpdk/spdk_pid72615 00:17:38.166 Removing: /var/run/dpdk/spdk_pid73216 00:17:38.166 Removing: /var/run/dpdk/spdk_pid73419 00:17:38.166 Removing: /var/run/dpdk/spdk_pid73549 00:17:38.166 Removing: /var/run/dpdk/spdk_pid74148 00:17:38.166 Removing: /var/run/dpdk/spdk_pid74462 00:17:38.166 Removing: /var/run/dpdk/spdk_pid74591 00:17:38.166 Removing: /var/run/dpdk/spdk_pid75926 00:17:38.166 Removing: /var/run/dpdk/spdk_pid76163 00:17:38.166 Removing: /var/run/dpdk/spdk_pid76292 00:17:38.166 Removing: /var/run/dpdk/spdk_pid77527 00:17:38.166 Removing: /var/run/dpdk/spdk_pid77740 00:17:38.166 Removing: /var/run/dpdk/spdk_pid77851 00:17:38.166 Removing: /var/run/dpdk/spdk_pid79066 00:17:38.166 Removing: /var/run/dpdk/spdk_pid79459 00:17:38.166 Removing: /var/run/dpdk/spdk_pid79570 00:17:38.166 Removing: /var/run/dpdk/spdk_pid80863 00:17:38.166 Removing: /var/run/dpdk/spdk_pid81086 00:17:38.166 Removing: /var/run/dpdk/spdk_pid81198 00:17:38.166 Removing: /var/run/dpdk/spdk_pid82486 00:17:38.166 Removing: /var/run/dpdk/spdk_pid82710 00:17:38.166 Removing: /var/run/dpdk/spdk_pid82822 00:17:38.166 Removing: /var/run/dpdk/spdk_pid84116 00:17:38.166 Removing: /var/run/dpdk/spdk_pid84545 00:17:38.166 Removing: /var/run/dpdk/spdk_pid84657 00:17:38.166 Removing: /var/run/dpdk/spdk_pid84766 00:17:38.166 Removing: /var/run/dpdk/spdk_pid85090 00:17:38.166 Removing: /var/run/dpdk/spdk_pid85673 00:17:38.166 Removing: /var/run/dpdk/spdk_pid85989 00:17:38.166 Removing: /var/run/dpdk/spdk_pid86552 00:17:38.166 Removing: /var/run/dpdk/spdk_pid86890 00:17:38.166 Removing: /var/run/dpdk/spdk_pid87494 00:17:38.166 Removing: /var/run/dpdk/spdk_pid87820 00:17:38.166 Removing: /var/run/dpdk/spdk_pid89513 00:17:38.166 Removing: /var/run/dpdk/spdk_pid89904 00:17:38.166 Removing: /var/run/dpdk/spdk_pid90244 00:17:38.166 Removing: /var/run/dpdk/spdk_pid92042 00:17:38.166 Removing: /var/run/dpdk/spdk_pid92470 00:17:38.166 Removing: /var/run/dpdk/spdk_pid92849 00:17:38.166 Removing: /var/run/dpdk/spdk_pid93714 00:17:38.166 Removing: /var/run/dpdk/spdk_pid93996 00:17:38.166 Removing: /var/run/dpdk/spdk_pid94780 00:17:38.166 Removing: /var/run/dpdk/spdk_pid95067 00:17:38.166 Removing: /var/run/dpdk/spdk_pid95851 00:17:38.166 Removing: /var/run/dpdk/spdk_pid96137 00:17:38.166 Removing: /var/run/dpdk/spdk_pid96709 00:17:38.166 Removing: /var/run/dpdk/spdk_pid96921 00:17:38.166 Removing: /var/run/dpdk/spdk_pid96954 00:17:38.166 Removing: /var/run/dpdk/spdk_pid96974 00:17:38.166 Removing: /var/run/dpdk/spdk_pid97171 00:17:38.166 Removing: /var/run/dpdk/spdk_pid97270 00:17:38.166 Removing: /var/run/dpdk/spdk_pid97316 00:17:38.166 Removing: /var/run/dpdk/spdk_pid97363 00:17:38.425 Removing: /var/run/dpdk/spdk_pid97392 00:17:38.425 Removing: /var/run/dpdk/spdk_pid97412 00:17:38.425 Clean 00:17:38.425 19:07:55 -- common/autotest_common.sh@1453 -- # return 0 00:17:38.425 19:07:55 -- spdk/autotest.sh@389 -- # timing_exit post_cleanup 00:17:38.425 19:07:55 -- common/autotest_common.sh@732 -- # xtrace_disable 00:17:38.425 19:07:55 -- common/autotest_common.sh@10 -- # set +x 00:17:38.425 19:07:55 -- spdk/autotest.sh@391 -- # timing_exit autotest 00:17:38.425 19:07:55 -- common/autotest_common.sh@732 -- # xtrace_disable 00:17:38.425 19:07:55 -- common/autotest_common.sh@10 -- # set +x 00:17:38.425 19:07:55 -- spdk/autotest.sh@392 -- # chmod a+r /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:17:38.425 19:07:55 -- spdk/autotest.sh@394 -- # [[ -f /home/vagrant/spdk_repo/spdk/../output/udev.log ]] 00:17:38.425 19:07:55 -- spdk/autotest.sh@394 -- # rm -f /home/vagrant/spdk_repo/spdk/../output/udev.log 00:17:38.425 19:07:55 -- spdk/autotest.sh@396 -- # [[ y == y ]] 00:17:38.425 19:07:55 -- spdk/autotest.sh@398 -- # hostname 00:17:38.425 19:07:55 -- spdk/autotest.sh@398 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /home/vagrant/spdk_repo/spdk -t fedora39-cloud-1721788873-2326 -o /home/vagrant/spdk_repo/spdk/../output/cov_test.info 00:17:38.685 geninfo: WARNING: invalid characters removed from testname! 00:18:00.645 19:08:17 -- spdk/autotest.sh@399 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /home/vagrant/spdk_repo/spdk/../output/cov_base.info -a /home/vagrant/spdk_repo/spdk/../output/cov_test.info -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:03.188 19:08:20 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/dpdk/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:05.103 19:08:22 -- spdk/autotest.sh@404 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:07.014 19:08:24 -- spdk/autotest.sh@405 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/examples/vmd/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:09.554 19:08:26 -- spdk/autotest.sh@406 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:11.460 19:08:28 -- spdk/autotest.sh@407 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:13.999 19:08:31 -- spdk/autotest.sh@408 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:18:13.999 19:08:31 -- spdk/autorun.sh@1 -- $ timing_finish 00:18:13.999 19:08:31 -- common/autotest_common.sh@738 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/timing.txt ]] 00:18:13.999 19:08:31 -- common/autotest_common.sh@740 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:18:13.999 19:08:31 -- common/autotest_common.sh@741 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:18:13.999 19:08:31 -- common/autotest_common.sh@744 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:18:13.999 + [[ -n 6157 ]] 00:18:13.999 + sudo kill 6157 00:18:14.016 [Pipeline] } 00:18:14.055 [Pipeline] // timeout 00:18:14.059 [Pipeline] } 00:18:14.074 [Pipeline] // stage 00:18:14.077 [Pipeline] } 00:18:14.090 [Pipeline] // catchError 00:18:14.098 [Pipeline] stage 00:18:14.100 [Pipeline] { (Stop VM) 00:18:14.112 [Pipeline] sh 00:18:14.396 + vagrant halt 00:18:16.937 ==> default: Halting domain... 00:18:25.109 [Pipeline] sh 00:18:25.398 + vagrant destroy -f 00:18:27.940 ==> default: Removing domain... 00:18:27.952 [Pipeline] sh 00:18:28.236 + mv output /var/jenkins/workspace/raid-vg-autotest/output 00:18:28.246 [Pipeline] } 00:18:28.263 [Pipeline] // stage 00:18:28.268 [Pipeline] } 00:18:28.282 [Pipeline] // dir 00:18:28.288 [Pipeline] } 00:18:28.303 [Pipeline] // wrap 00:18:28.309 [Pipeline] } 00:18:28.325 [Pipeline] // catchError 00:18:28.336 [Pipeline] stage 00:18:28.339 [Pipeline] { (Epilogue) 00:18:28.355 [Pipeline] sh 00:18:28.645 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:18:32.859 [Pipeline] catchError 00:18:32.861 [Pipeline] { 00:18:32.872 [Pipeline] sh 00:18:33.157 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:18:33.157 Artifacts sizes are good 00:18:33.166 [Pipeline] } 00:18:33.182 [Pipeline] // catchError 00:18:33.192 [Pipeline] archiveArtifacts 00:18:33.199 Archiving artifacts 00:18:33.316 [Pipeline] cleanWs 00:18:33.327 [WS-CLEANUP] Deleting project workspace... 00:18:33.327 [WS-CLEANUP] Deferred wipeout is used... 00:18:33.333 [WS-CLEANUP] done 00:18:33.335 [Pipeline] } 00:18:33.344 [Pipeline] // stage 00:18:33.347 [Pipeline] } 00:18:33.354 [Pipeline] // node 00:18:33.357 [Pipeline] End of Pipeline 00:18:33.452 Finished: SUCCESS